var/home/core/zuul-output/0000755000175000017500000000000015111107152014517 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111120226015460 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005052250415111120217017667 0ustar rootrootNov 24 17:00:40 crc systemd[1]: Starting Kubernetes Kubelet... Nov 24 17:00:40 crc restorecon[4678]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:00:40 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:00:41 crc restorecon[4678]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 17:00:41 crc restorecon[4678]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 24 17:00:42 crc kubenswrapper[4777]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 17:00:42 crc kubenswrapper[4777]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 24 17:00:42 crc kubenswrapper[4777]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 17:00:42 crc kubenswrapper[4777]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 17:00:42 crc kubenswrapper[4777]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 24 17:00:42 crc kubenswrapper[4777]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.834165 4777 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839626 4777 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839658 4777 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839667 4777 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839677 4777 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839687 4777 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839697 4777 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839706 4777 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839714 4777 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839722 4777 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839731 4777 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839739 4777 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839747 4777 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839755 4777 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839763 4777 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839770 4777 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839778 4777 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839785 4777 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839807 4777 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839816 4777 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839824 4777 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839832 4777 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839840 4777 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839847 4777 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839858 4777 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839869 4777 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839878 4777 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839887 4777 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839894 4777 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839902 4777 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839910 4777 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839918 4777 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839929 4777 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839940 4777 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839950 4777 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839960 4777 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.839993 4777 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840002 4777 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840010 4777 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840018 4777 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840029 4777 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840038 4777 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840046 4777 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840054 4777 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840064 4777 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840072 4777 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840081 4777 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840089 4777 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840097 4777 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840106 4777 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840114 4777 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840122 4777 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840130 4777 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840138 4777 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840148 4777 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840160 4777 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840170 4777 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840178 4777 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840187 4777 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840195 4777 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840202 4777 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840210 4777 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840218 4777 feature_gate.go:330] unrecognized feature gate: Example Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840226 4777 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840233 4777 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840240 4777 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840249 4777 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840256 4777 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840264 4777 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840271 4777 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840278 4777 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.840286 4777 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841075 4777 flags.go:64] FLAG: --address="0.0.0.0" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841097 4777 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841110 4777 flags.go:64] FLAG: --anonymous-auth="true" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841121 4777 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841132 4777 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841142 4777 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841153 4777 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841164 4777 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841173 4777 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841182 4777 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841193 4777 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841202 4777 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841211 4777 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841219 4777 flags.go:64] FLAG: --cgroup-root="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841228 4777 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841237 4777 flags.go:64] FLAG: --client-ca-file="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841247 4777 flags.go:64] FLAG: --cloud-config="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841255 4777 flags.go:64] FLAG: --cloud-provider="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841267 4777 flags.go:64] FLAG: --cluster-dns="[]" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841278 4777 flags.go:64] FLAG: --cluster-domain="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841287 4777 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841296 4777 flags.go:64] FLAG: --config-dir="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841305 4777 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841314 4777 flags.go:64] FLAG: --container-log-max-files="5" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841324 4777 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841334 4777 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841343 4777 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841352 4777 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841361 4777 flags.go:64] FLAG: --contention-profiling="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841370 4777 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841379 4777 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841388 4777 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841397 4777 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841408 4777 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841417 4777 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841426 4777 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841435 4777 flags.go:64] FLAG: --enable-load-reader="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841444 4777 flags.go:64] FLAG: --enable-server="true" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841453 4777 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841464 4777 flags.go:64] FLAG: --event-burst="100" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841473 4777 flags.go:64] FLAG: --event-qps="50" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841482 4777 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841491 4777 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841500 4777 flags.go:64] FLAG: --eviction-hard="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841511 4777 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841519 4777 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841529 4777 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841538 4777 flags.go:64] FLAG: --eviction-soft="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841553 4777 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841562 4777 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841571 4777 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841579 4777 flags.go:64] FLAG: --experimental-mounter-path="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841588 4777 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841597 4777 flags.go:64] FLAG: --fail-swap-on="true" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841607 4777 flags.go:64] FLAG: --feature-gates="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841617 4777 flags.go:64] FLAG: --file-check-frequency="20s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841626 4777 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841635 4777 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841644 4777 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841653 4777 flags.go:64] FLAG: --healthz-port="10248" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841662 4777 flags.go:64] FLAG: --help="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841671 4777 flags.go:64] FLAG: --hostname-override="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841680 4777 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841688 4777 flags.go:64] FLAG: --http-check-frequency="20s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841697 4777 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841706 4777 flags.go:64] FLAG: --image-credential-provider-config="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841714 4777 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841723 4777 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841731 4777 flags.go:64] FLAG: --image-service-endpoint="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841740 4777 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841748 4777 flags.go:64] FLAG: --kube-api-burst="100" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841757 4777 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841766 4777 flags.go:64] FLAG: --kube-api-qps="50" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841776 4777 flags.go:64] FLAG: --kube-reserved="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841785 4777 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841794 4777 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841803 4777 flags.go:64] FLAG: --kubelet-cgroups="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841812 4777 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841821 4777 flags.go:64] FLAG: --lock-file="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841830 4777 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841842 4777 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841851 4777 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841867 4777 flags.go:64] FLAG: --log-json-split-stream="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841877 4777 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841887 4777 flags.go:64] FLAG: --log-text-split-stream="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841896 4777 flags.go:64] FLAG: --logging-format="text" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841904 4777 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841914 4777 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841922 4777 flags.go:64] FLAG: --manifest-url="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841931 4777 flags.go:64] FLAG: --manifest-url-header="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841943 4777 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841952 4777 flags.go:64] FLAG: --max-open-files="1000000" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841963 4777 flags.go:64] FLAG: --max-pods="110" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.841995 4777 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842005 4777 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842014 4777 flags.go:64] FLAG: --memory-manager-policy="None" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842023 4777 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842032 4777 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842041 4777 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842050 4777 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842068 4777 flags.go:64] FLAG: --node-status-max-images="50" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842077 4777 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842086 4777 flags.go:64] FLAG: --oom-score-adj="-999" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842095 4777 flags.go:64] FLAG: --pod-cidr="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842104 4777 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842117 4777 flags.go:64] FLAG: --pod-manifest-path="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842151 4777 flags.go:64] FLAG: --pod-max-pids="-1" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842161 4777 flags.go:64] FLAG: --pods-per-core="0" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842170 4777 flags.go:64] FLAG: --port="10250" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842179 4777 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842188 4777 flags.go:64] FLAG: --provider-id="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842197 4777 flags.go:64] FLAG: --qos-reserved="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842209 4777 flags.go:64] FLAG: --read-only-port="10255" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842218 4777 flags.go:64] FLAG: --register-node="true" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842227 4777 flags.go:64] FLAG: --register-schedulable="true" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842236 4777 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842250 4777 flags.go:64] FLAG: --registry-burst="10" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842259 4777 flags.go:64] FLAG: --registry-qps="5" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842268 4777 flags.go:64] FLAG: --reserved-cpus="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842277 4777 flags.go:64] FLAG: --reserved-memory="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842288 4777 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842297 4777 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842306 4777 flags.go:64] FLAG: --rotate-certificates="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842314 4777 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842323 4777 flags.go:64] FLAG: --runonce="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842332 4777 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842342 4777 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842351 4777 flags.go:64] FLAG: --seccomp-default="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842360 4777 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842369 4777 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842378 4777 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842387 4777 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842396 4777 flags.go:64] FLAG: --storage-driver-password="root" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842404 4777 flags.go:64] FLAG: --storage-driver-secure="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842413 4777 flags.go:64] FLAG: --storage-driver-table="stats" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842422 4777 flags.go:64] FLAG: --storage-driver-user="root" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842431 4777 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842440 4777 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842448 4777 flags.go:64] FLAG: --system-cgroups="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842457 4777 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842471 4777 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842480 4777 flags.go:64] FLAG: --tls-cert-file="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842488 4777 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842499 4777 flags.go:64] FLAG: --tls-min-version="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842511 4777 flags.go:64] FLAG: --tls-private-key-file="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842520 4777 flags.go:64] FLAG: --topology-manager-policy="none" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842529 4777 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842538 4777 flags.go:64] FLAG: --topology-manager-scope="container" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842547 4777 flags.go:64] FLAG: --v="2" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842558 4777 flags.go:64] FLAG: --version="false" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842570 4777 flags.go:64] FLAG: --vmodule="" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842580 4777 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.842590 4777 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843289 4777 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843305 4777 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843314 4777 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843323 4777 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843331 4777 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843340 4777 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843348 4777 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843356 4777 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843364 4777 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843374 4777 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843382 4777 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843390 4777 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843397 4777 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843405 4777 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843415 4777 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843425 4777 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843435 4777 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843444 4777 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843452 4777 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843461 4777 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843469 4777 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843479 4777 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843489 4777 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843502 4777 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843512 4777 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843520 4777 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843528 4777 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843536 4777 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843545 4777 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843553 4777 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843560 4777 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843569 4777 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843576 4777 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843584 4777 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843592 4777 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843599 4777 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843607 4777 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843614 4777 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843622 4777 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843630 4777 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843637 4777 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843645 4777 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843653 4777 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843660 4777 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843668 4777 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843679 4777 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843689 4777 feature_gate.go:330] unrecognized feature gate: Example Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843698 4777 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843706 4777 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843714 4777 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843722 4777 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843730 4777 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843738 4777 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843746 4777 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843754 4777 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843764 4777 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843773 4777 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843783 4777 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843791 4777 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843798 4777 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843806 4777 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843814 4777 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843822 4777 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843830 4777 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843838 4777 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843847 4777 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843855 4777 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843863 4777 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843873 4777 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843881 4777 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.843890 4777 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.844537 4777 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.857434 4777 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.857498 4777 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857624 4777 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857644 4777 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857654 4777 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857666 4777 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857677 4777 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857686 4777 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857696 4777 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857704 4777 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857712 4777 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857720 4777 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857728 4777 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857735 4777 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857743 4777 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857754 4777 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857767 4777 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857775 4777 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857784 4777 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857795 4777 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857802 4777 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857810 4777 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857818 4777 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857827 4777 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857836 4777 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857845 4777 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857854 4777 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857862 4777 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857870 4777 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857878 4777 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857885 4777 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857893 4777 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857900 4777 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857908 4777 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857916 4777 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857924 4777 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857932 4777 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857939 4777 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857947 4777 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857955 4777 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857963 4777 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.857996 4777 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858003 4777 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858011 4777 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858018 4777 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858026 4777 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858034 4777 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858042 4777 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858051 4777 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858058 4777 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858066 4777 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858074 4777 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858082 4777 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858089 4777 feature_gate.go:330] unrecognized feature gate: Example Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858097 4777 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858104 4777 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858116 4777 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858125 4777 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858135 4777 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858144 4777 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858154 4777 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858165 4777 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858192 4777 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858200 4777 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858209 4777 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858217 4777 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858226 4777 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858234 4777 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858242 4777 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858250 4777 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858258 4777 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858267 4777 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858275 4777 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.858288 4777 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858577 4777 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858616 4777 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858626 4777 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858637 4777 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858649 4777 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858657 4777 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858665 4777 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858674 4777 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858682 4777 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858690 4777 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858698 4777 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858705 4777 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858713 4777 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858721 4777 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858729 4777 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858737 4777 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858746 4777 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858755 4777 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858763 4777 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858773 4777 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858783 4777 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858792 4777 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858801 4777 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858810 4777 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858818 4777 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858826 4777 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858834 4777 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858842 4777 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858850 4777 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858857 4777 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858865 4777 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858873 4777 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858880 4777 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858888 4777 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858895 4777 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858903 4777 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858911 4777 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858919 4777 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858927 4777 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858936 4777 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858943 4777 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858951 4777 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858959 4777 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858988 4777 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.858997 4777 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859005 4777 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859012 4777 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859020 4777 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859028 4777 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859038 4777 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859045 4777 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859054 4777 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859061 4777 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859069 4777 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859077 4777 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859084 4777 feature_gate.go:330] unrecognized feature gate: Example Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859092 4777 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859099 4777 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859107 4777 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859116 4777 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859123 4777 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859133 4777 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859143 4777 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859152 4777 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859160 4777 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859168 4777 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859176 4777 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859184 4777 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859193 4777 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859200 4777 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 17:00:42 crc kubenswrapper[4777]: W1124 17:00:42.859208 4777 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.859220 4777 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.860995 4777 server.go:940] "Client rotation is on, will bootstrap in background" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.867830 4777 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.868008 4777 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.872140 4777 server.go:997] "Starting client certificate rotation" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.872190 4777 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.873961 4777 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-18 05:18:33.488087962 +0000 UTC Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.874105 4777 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1308h17m50.613988159s for next certificate rotation Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.946098 4777 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.952937 4777 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 17:00:42 crc kubenswrapper[4777]: I1124 17:00:42.987096 4777 log.go:25] "Validated CRI v1 runtime API" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.082301 4777 log.go:25] "Validated CRI v1 image API" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.084916 4777 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.091680 4777 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-24-16-56-05-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.091724 4777 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:45 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.116195 4777 manager.go:217] Machine: {Timestamp:2025-11-24 17:00:43.111110177 +0000 UTC m=+1.270305246 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:ab3e815f-373b-46b5-8331-bc00cff3390c BootID:93f6ada4-92ef-4dea-97e3-58231cc70ccf Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:45 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:50:90:91 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:50:90:91 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:f7:dc:dd Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:6d:ca:14 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:48:51:78 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:e3:df:93 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:da:28:7d:b9:09:54 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:06:83:e6:44:ff:0c Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.116450 4777 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.116707 4777 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.117138 4777 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.117292 4777 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.117329 4777 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.117511 4777 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.117520 4777 container_manager_linux.go:303] "Creating device plugin manager" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.119371 4777 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.120030 4777 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.120191 4777 state_mem.go:36] "Initialized new in-memory state store" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.120270 4777 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.127318 4777 kubelet.go:418] "Attempting to sync node with API server" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.127348 4777 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.127496 4777 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.127514 4777 kubelet.go:324] "Adding apiserver pod source" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.127554 4777 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.143360 4777 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 24 17:00:43 crc kubenswrapper[4777]: W1124 17:00:43.145899 4777 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:43 crc kubenswrapper[4777]: W1124 17:00:43.145891 4777 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.146049 4777 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.146065 4777 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.149728 4777 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.152416 4777 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.155901 4777 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.155932 4777 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.155943 4777 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.155953 4777 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.155994 4777 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.156005 4777 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.156018 4777 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.156039 4777 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.156056 4777 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.156068 4777 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.156082 4777 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.156091 4777 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.160372 4777 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.161272 4777 server.go:1280] "Started kubelet" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.161542 4777 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.161751 4777 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.162298 4777 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 17:00:43 crc systemd[1]: Started Kubernetes Kubelet. Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.164847 4777 server.go:460] "Adding debug handlers to kubelet server" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.165256 4777 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.165320 4777 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.165341 4777 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.165412 4777 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 20:07:37.063149099 +0000 UTC Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.165469 4777 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 507h6m53.89768378s for next certificate rotation Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.165651 4777 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.165677 4777 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.165786 4777 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.166058 4777 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 17:00:43 crc kubenswrapper[4777]: W1124 17:00:43.166612 4777 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.166688 4777 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.166709 4777 factory.go:55] Registering systemd factory Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.166740 4777 factory.go:221] Registration of the systemd container factory successfully Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.167784 4777 factory.go:153] Registering CRI-O factory Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.167866 4777 factory.go:221] Registration of the crio container factory successfully Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.167960 4777 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.168049 4777 factory.go:103] Registering Raw factory Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.168074 4777 manager.go:1196] Started watching for new ooms in manager Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.168730 4777 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="200ms" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.172137 4777 manager.go:319] Starting recovery of all containers Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.182923 4777 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.9:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187afff7a51e934d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 17:00:43.161219917 +0000 UTC m=+1.320414996,LastTimestamp:2025-11-24 17:00:43.161219917 +0000 UTC m=+1.320414996,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195037 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195129 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195145 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195159 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195175 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195189 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195203 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195218 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195237 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195255 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195272 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195288 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195302 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195876 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195918 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195928 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195938 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195950 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.195961 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196023 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196037 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196049 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196059 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196070 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196080 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196089 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196134 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196146 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196157 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196168 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196180 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196211 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196224 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196235 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196244 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196256 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196265 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196276 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196287 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196298 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196309 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196321 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196333 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196343 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196355 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196364 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196378 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196390 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196404 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196416 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196428 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196441 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196455 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196467 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196478 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196512 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196523 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196533 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196543 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196553 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196562 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196573 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196584 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196594 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196605 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196614 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196622 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196636 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196646 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196656 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196666 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196676 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196686 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196696 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196708 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196718 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196728 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196738 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196788 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196805 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196817 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196829 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196841 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196850 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196859 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196886 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196896 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196905 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196915 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196925 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196935 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196945 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196956 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.196980 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197013 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197024 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197035 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197046 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197058 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197073 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197084 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197100 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197113 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197127 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197142 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197154 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197164 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197195 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197209 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197219 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197233 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197244 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197255 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197267 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197279 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197290 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197299 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197310 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197320 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197331 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197343 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197353 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197362 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197372 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197382 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197392 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197403 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197412 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197422 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197432 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197442 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197457 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197467 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197478 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197488 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197497 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197507 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197519 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197531 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197541 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197551 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197561 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197573 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197583 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197593 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197603 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197614 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197624 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197639 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197650 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197675 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197687 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197699 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197710 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197722 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197737 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197748 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197760 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197772 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197784 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197797 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197808 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197820 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.197838 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.202890 4777 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.202935 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.202951 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.202962 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.202995 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203008 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203039 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203052 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203064 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203077 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203094 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203108 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203119 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203136 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203147 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203157 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203168 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203178 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203189 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203200 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203210 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203220 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203229 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203240 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203251 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203266 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203279 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203289 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203299 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203311 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203321 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203334 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203344 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203356 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203365 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203376 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203389 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203399 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203409 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203420 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203431 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203442 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203454 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203465 4777 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203474 4777 reconstruct.go:97] "Volume reconstruction finished" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.203482 4777 reconciler.go:26] "Reconciler: start to sync state" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.211425 4777 manager.go:324] Recovery completed Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.229800 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.233300 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.233355 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.233375 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.235354 4777 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.235376 4777 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.235411 4777 state_mem.go:36] "Initialized new in-memory state store" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.240525 4777 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.243783 4777 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.243840 4777 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.243888 4777 kubelet.go:2335] "Starting kubelet main sync loop" Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.243999 4777 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 17:00:43 crc kubenswrapper[4777]: W1124 17:00:43.245029 4777 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.245128 4777 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.266160 4777 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.315476 4777 policy_none.go:49] "None policy: Start" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.317106 4777 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.317204 4777 state_mem.go:35] "Initializing new in-memory state store" Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.344532 4777 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.366297 4777 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.370138 4777 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="400ms" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.389631 4777 manager.go:334] "Starting Device Plugin manager" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.389714 4777 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.389732 4777 server.go:79] "Starting device plugin registration server" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.390480 4777 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.390523 4777 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.390871 4777 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.390954 4777 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.390961 4777 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.402926 4777 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.491531 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.493392 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.493468 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.493510 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.493588 4777 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.494512 4777 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.544770 4777 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.544859 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.546268 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.546322 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.546339 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.546564 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.546808 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.546839 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.548018 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.548059 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.548100 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.548742 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.548774 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.548788 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.548894 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.549726 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.549819 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.552368 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.552417 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.552439 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.553118 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.553304 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.553336 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.554819 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.555072 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.555147 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.556016 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.556055 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.556069 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.556210 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.556437 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.556506 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.556580 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.556621 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.556639 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.557083 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.557126 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.557146 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.557380 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.557430 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.557942 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.558000 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.558012 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.558323 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.558369 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.558389 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.609343 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.609403 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.609436 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.609462 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.609484 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.609508 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.609533 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.609676 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.609819 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.609877 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.609934 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.610028 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.610079 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.610110 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.610130 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.695110 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.696507 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.696558 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.696575 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.696607 4777 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.697102 4777 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.711521 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.711591 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.711644 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.711678 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.711714 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.711654 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.711673 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.711839 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.711841 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.711761 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.711934 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712000 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712008 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712034 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712054 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712066 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712097 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712129 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712135 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712159 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712215 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712226 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712259 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712296 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712270 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712272 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712333 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712423 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712456 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.712506 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.770891 4777 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="800ms" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.890819 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.909956 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.937141 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.946268 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: I1124 17:00:43.949423 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:00:43 crc kubenswrapper[4777]: W1124 17:00:43.990140 4777 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:43 crc kubenswrapper[4777]: E1124 17:00:43.990258 4777 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:00:44 crc kubenswrapper[4777]: W1124 17:00:44.057553 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-4520fc19ddab32d0739161aafb748eb86c61d998895a30caa7a3f035a02d536b WatchSource:0}: Error finding container 4520fc19ddab32d0739161aafb748eb86c61d998895a30caa7a3f035a02d536b: Status 404 returned error can't find the container with id 4520fc19ddab32d0739161aafb748eb86c61d998895a30caa7a3f035a02d536b Nov 24 17:00:44 crc kubenswrapper[4777]: W1124 17:00:44.063112 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-14c460d61a1a243af5fe28773ec87aabc81f06511a4faac01adced199d88049c WatchSource:0}: Error finding container 14c460d61a1a243af5fe28773ec87aabc81f06511a4faac01adced199d88049c: Status 404 returned error can't find the container with id 14c460d61a1a243af5fe28773ec87aabc81f06511a4faac01adced199d88049c Nov 24 17:00:44 crc kubenswrapper[4777]: W1124 17:00:44.065910 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-c6380f4b3a7e11804bb1a7128f87194a62559a9476248fb234bdfa42a5c335a0 WatchSource:0}: Error finding container c6380f4b3a7e11804bb1a7128f87194a62559a9476248fb234bdfa42a5c335a0: Status 404 returned error can't find the container with id c6380f4b3a7e11804bb1a7128f87194a62559a9476248fb234bdfa42a5c335a0 Nov 24 17:00:44 crc kubenswrapper[4777]: W1124 17:00:44.068043 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-4613b52ac77b1e0c25e5ce884e58ac8ad3f787296d09438c0977546d483092cf WatchSource:0}: Error finding container 4613b52ac77b1e0c25e5ce884e58ac8ad3f787296d09438c0977546d483092cf: Status 404 returned error can't find the container with id 4613b52ac77b1e0c25e5ce884e58ac8ad3f787296d09438c0977546d483092cf Nov 24 17:00:44 crc kubenswrapper[4777]: W1124 17:00:44.074451 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-acad28c404cbc7f5d1ee199a2b00b24b2727e302e8236b090121bcf3a1584bc0 WatchSource:0}: Error finding container acad28c404cbc7f5d1ee199a2b00b24b2727e302e8236b090121bcf3a1584bc0: Status 404 returned error can't find the container with id acad28c404cbc7f5d1ee199a2b00b24b2727e302e8236b090121bcf3a1584bc0 Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.097765 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.099248 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.099349 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.099369 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.099443 4777 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 17:00:44 crc kubenswrapper[4777]: E1124 17:00:44.100389 4777 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 24 17:00:44 crc kubenswrapper[4777]: W1124 17:00:44.106123 4777 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:44 crc kubenswrapper[4777]: E1124 17:00:44.106231 4777 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:00:44 crc kubenswrapper[4777]: W1124 17:00:44.127757 4777 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:44 crc kubenswrapper[4777]: E1124 17:00:44.128312 4777 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.172648 4777 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.251054 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"acad28c404cbc7f5d1ee199a2b00b24b2727e302e8236b090121bcf3a1584bc0"} Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.253169 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4613b52ac77b1e0c25e5ce884e58ac8ad3f787296d09438c0977546d483092cf"} Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.254887 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c6380f4b3a7e11804bb1a7128f87194a62559a9476248fb234bdfa42a5c335a0"} Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.255902 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"14c460d61a1a243af5fe28773ec87aabc81f06511a4faac01adced199d88049c"} Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.257066 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4520fc19ddab32d0739161aafb748eb86c61d998895a30caa7a3f035a02d536b"} Nov 24 17:00:44 crc kubenswrapper[4777]: E1124 17:00:44.572483 4777 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="1.6s" Nov 24 17:00:44 crc kubenswrapper[4777]: W1124 17:00:44.636909 4777 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:44 crc kubenswrapper[4777]: E1124 17:00:44.637052 4777 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.901287 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.904120 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.904191 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.904215 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:44 crc kubenswrapper[4777]: I1124 17:00:44.904263 4777 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 17:00:44 crc kubenswrapper[4777]: E1124 17:00:44.905128 4777 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 24 17:00:45 crc kubenswrapper[4777]: I1124 17:00:45.166629 4777 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.167125 4777 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:46 crc kubenswrapper[4777]: E1124 17:00:46.173692 4777 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="3.2s" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.265484 4777 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7" exitCode=0 Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.265584 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7"} Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.265616 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.266686 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.266718 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.266729 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.268308 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.269066 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.269097 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.269108 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.269200 4777 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="104e9b5dc0c02e1ea12feeb30c7d491adc0f671b881f68a0ce4e15304554b207" exitCode=0 Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.269273 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"104e9b5dc0c02e1ea12feeb30c7d491adc0f671b881f68a0ce4e15304554b207"} Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.269379 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.270856 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.270909 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.270927 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.271603 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9a551cc061d7105b0d4a9978872e1756073cc47290c5ed958f003ac770a2498f"} Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.271659 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"034f342bd386d862217b259fae5bdf49b567ebbd9162b932f7d9b9f469a4e7c1"} Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.273889 4777 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f" exitCode=0 Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.273962 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f"} Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.274024 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.274912 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.274960 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.275008 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.276162 4777 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb" exitCode=0 Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.276205 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb"} Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.276266 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.277364 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.277404 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.277417 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:46 crc kubenswrapper[4777]: W1124 17:00:46.387164 4777 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:46 crc kubenswrapper[4777]: E1124 17:00:46.387275 4777 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:00:46 crc kubenswrapper[4777]: W1124 17:00:46.463395 4777 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:46 crc kubenswrapper[4777]: E1124 17:00:46.463476 4777 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.505551 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.506730 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.506764 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.506774 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:46 crc kubenswrapper[4777]: I1124 17:00:46.506799 4777 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 17:00:46 crc kubenswrapper[4777]: E1124 17:00:46.507276 4777 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 24 17:00:46 crc kubenswrapper[4777]: W1124 17:00:46.720041 4777 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:46 crc kubenswrapper[4777]: E1124 17:00:46.720163 4777 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:00:46 crc kubenswrapper[4777]: W1124 17:00:46.918889 4777 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:46 crc kubenswrapper[4777]: E1124 17:00:46.919046 4777 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.166741 4777 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.281658 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453"} Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.282072 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6"} Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.282095 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd"} Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.284832 4777 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="700f6f49ed9507d3c3074fc19424595e0dbc2d95b1a51a76cfa65a0b8642e68e" exitCode=0 Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.285071 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"700f6f49ed9507d3c3074fc19424595e0dbc2d95b1a51a76cfa65a0b8642e68e"} Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.285168 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.286691 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.286755 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.286780 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.290369 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6cdd6568224ddbf83698ac6b0e23ebc789f64d4d2090c64d80a4681d9bb70668"} Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.290468 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a9d0dc8c72b79ffb367d3a2120767ba87fa708f93aafa7950733826e935478ad"} Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.290399 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.295366 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.295367 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"13ad8466a0f0b306dd124ed16530743baeb94e28b3e078e854859aaa093edf73"} Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.296466 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.296526 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.296545 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.297626 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.297662 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.297676 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.300019 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211"} Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.300054 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d"} Nov 24 17:00:47 crc kubenswrapper[4777]: I1124 17:00:47.300068 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da"} Nov 24 17:00:47 crc kubenswrapper[4777]: E1124 17:00:47.927656 4777 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.9:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187afff7a51e934d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 17:00:43.161219917 +0000 UTC m=+1.320414996,LastTimestamp:2025-11-24 17:00:43.161219917 +0000 UTC m=+1.320414996,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.167542 4777 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.307826 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6967922795b8929633b76e4ac97d832cae0f3f53600631b7cb52d16690f58e98"} Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.307920 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633"} Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.311526 4777 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="55297085683d62c3b6c607487741e50eb2ac526c3304b140384bb33f85ce6d6b" exitCode=0 Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.311606 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"55297085683d62c3b6c607487741e50eb2ac526c3304b140384bb33f85ce6d6b"} Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.311688 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.311733 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.311691 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.311850 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.313610 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.313660 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.313687 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.313708 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.313686 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.313748 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.313862 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.313907 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.313928 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.314607 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.314659 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.314678 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:48 crc kubenswrapper[4777]: I1124 17:00:48.906815 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:00:49 crc kubenswrapper[4777]: I1124 17:00:49.167077 4777 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:49 crc kubenswrapper[4777]: I1124 17:00:49.319286 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bf8026bb6460f3b94c6e148f2f6e1cec97160b9b8ac26bd022f4a828faa5d38f"} Nov 24 17:00:49 crc kubenswrapper[4777]: I1124 17:00:49.319425 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:49 crc kubenswrapper[4777]: I1124 17:00:49.319552 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:49 crc kubenswrapper[4777]: I1124 17:00:49.321325 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:49 crc kubenswrapper[4777]: I1124 17:00:49.321393 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:49 crc kubenswrapper[4777]: I1124 17:00:49.321413 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:49 crc kubenswrapper[4777]: I1124 17:00:49.321442 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:49 crc kubenswrapper[4777]: I1124 17:00:49.321504 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:49 crc kubenswrapper[4777]: I1124 17:00:49.321517 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:49 crc kubenswrapper[4777]: E1124 17:00:49.375583 4777 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="6.4s" Nov 24 17:00:49 crc kubenswrapper[4777]: W1124 17:00:49.691088 4777 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:49 crc kubenswrapper[4777]: E1124 17:00:49.691186 4777 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 24 17:00:49 crc kubenswrapper[4777]: I1124 17:00:49.707428 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:49 crc kubenswrapper[4777]: I1124 17:00:49.709018 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:49 crc kubenswrapper[4777]: I1124 17:00:49.709067 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:49 crc kubenswrapper[4777]: I1124 17:00:49.709127 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:49 crc kubenswrapper[4777]: I1124 17:00:49.709175 4777 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 17:00:49 crc kubenswrapper[4777]: E1124 17:00:49.709522 4777 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 24 17:00:50 crc kubenswrapper[4777]: I1124 17:00:50.167069 4777 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 24 17:00:50 crc kubenswrapper[4777]: I1124 17:00:50.325258 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"356940adde9dbdf959d83687973a3e697156f8fe126c44a8597da63bfa75275d"} Nov 24 17:00:50 crc kubenswrapper[4777]: I1124 17:00:50.325332 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fbf4e41a8419e135180170ee0fccae2eeb52f5ddae2c2fc6363594ec31354fca"} Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.163572 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.163773 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.165163 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.165213 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.165232 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.329378 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.332270 4777 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6967922795b8929633b76e4ac97d832cae0f3f53600631b7cb52d16690f58e98" exitCode=255 Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.332383 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6967922795b8929633b76e4ac97d832cae0f3f53600631b7cb52d16690f58e98"} Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.332686 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.334212 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.334265 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.334285 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.335291 4777 scope.go:117] "RemoveContainer" containerID="6967922795b8929633b76e4ac97d832cae0f3f53600631b7cb52d16690f58e98" Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.337709 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d259e984ddb8f711944a5528f33f8c7e9a9f85e6cccc5e111804a85cf9ac642f"} Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.337771 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7bd39c4a740bea1727c583d538f8008b020bf44860750851fed34535ee611cb0"} Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.337842 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.339003 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.339049 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.339068 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:51 crc kubenswrapper[4777]: I1124 17:00:51.956848 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:00:52 crc kubenswrapper[4777]: I1124 17:00:52.344321 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 17:00:52 crc kubenswrapper[4777]: I1124 17:00:52.346888 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2"} Nov 24 17:00:52 crc kubenswrapper[4777]: I1124 17:00:52.347040 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:52 crc kubenswrapper[4777]: I1124 17:00:52.347138 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:52 crc kubenswrapper[4777]: I1124 17:00:52.348532 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:52 crc kubenswrapper[4777]: I1124 17:00:52.348581 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:52 crc kubenswrapper[4777]: I1124 17:00:52.348601 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:52 crc kubenswrapper[4777]: I1124 17:00:52.348925 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:52 crc kubenswrapper[4777]: I1124 17:00:52.348963 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:52 crc kubenswrapper[4777]: I1124 17:00:52.349016 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:53 crc kubenswrapper[4777]: I1124 17:00:53.086925 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:00:53 crc kubenswrapper[4777]: I1124 17:00:53.113413 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:00:53 crc kubenswrapper[4777]: I1124 17:00:53.350375 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:53 crc kubenswrapper[4777]: I1124 17:00:53.350510 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:00:53 crc kubenswrapper[4777]: I1124 17:00:53.352823 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:53 crc kubenswrapper[4777]: I1124 17:00:53.352876 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:53 crc kubenswrapper[4777]: I1124 17:00:53.352895 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:53 crc kubenswrapper[4777]: E1124 17:00:53.403054 4777 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 17:00:54 crc kubenswrapper[4777]: I1124 17:00:54.353656 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:54 crc kubenswrapper[4777]: I1124 17:00:54.354949 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:54 crc kubenswrapper[4777]: I1124 17:00:54.355055 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:54 crc kubenswrapper[4777]: I1124 17:00:54.355128 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:54 crc kubenswrapper[4777]: I1124 17:00:54.450320 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 24 17:00:54 crc kubenswrapper[4777]: I1124 17:00:54.450874 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:54 crc kubenswrapper[4777]: I1124 17:00:54.452111 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:54 crc kubenswrapper[4777]: I1124 17:00:54.452154 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:54 crc kubenswrapper[4777]: I1124 17:00:54.452166 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:54 crc kubenswrapper[4777]: I1124 17:00:54.604695 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:00:54 crc kubenswrapper[4777]: I1124 17:00:54.604887 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:54 crc kubenswrapper[4777]: I1124 17:00:54.606358 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:54 crc kubenswrapper[4777]: I1124 17:00:54.606622 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:54 crc kubenswrapper[4777]: I1124 17:00:54.606796 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:55 crc kubenswrapper[4777]: I1124 17:00:55.450269 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:00:55 crc kubenswrapper[4777]: I1124 17:00:55.450582 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:55 crc kubenswrapper[4777]: I1124 17:00:55.452308 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:55 crc kubenswrapper[4777]: I1124 17:00:55.452373 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:55 crc kubenswrapper[4777]: I1124 17:00:55.452393 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:55 crc kubenswrapper[4777]: I1124 17:00:55.889214 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:00:55 crc kubenswrapper[4777]: I1124 17:00:55.889505 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:55 crc kubenswrapper[4777]: I1124 17:00:55.891249 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:55 crc kubenswrapper[4777]: I1124 17:00:55.891302 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:55 crc kubenswrapper[4777]: I1124 17:00:55.891321 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:55 crc kubenswrapper[4777]: I1124 17:00:55.897138 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:00:56 crc kubenswrapper[4777]: I1124 17:00:56.110480 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:56 crc kubenswrapper[4777]: I1124 17:00:56.112185 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:56 crc kubenswrapper[4777]: I1124 17:00:56.112239 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:56 crc kubenswrapper[4777]: I1124 17:00:56.112261 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:56 crc kubenswrapper[4777]: I1124 17:00:56.112295 4777 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 17:00:56 crc kubenswrapper[4777]: I1124 17:00:56.358797 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:56 crc kubenswrapper[4777]: I1124 17:00:56.360565 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:56 crc kubenswrapper[4777]: I1124 17:00:56.360617 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:56 crc kubenswrapper[4777]: I1124 17:00:56.360635 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:56 crc kubenswrapper[4777]: I1124 17:00:56.365325 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:00:57 crc kubenswrapper[4777]: I1124 17:00:57.363326 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:00:57 crc kubenswrapper[4777]: I1124 17:00:57.364744 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:00:57 crc kubenswrapper[4777]: I1124 17:00:57.364810 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:00:57 crc kubenswrapper[4777]: I1124 17:00:57.364836 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:00:57 crc kubenswrapper[4777]: I1124 17:00:57.605215 4777 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 17:00:57 crc kubenswrapper[4777]: I1124 17:00:57.605299 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 17:01:00 crc kubenswrapper[4777]: I1124 17:01:00.384060 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 24 17:01:00 crc kubenswrapper[4777]: I1124 17:01:00.384341 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:01:00 crc kubenswrapper[4777]: I1124 17:01:00.385724 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:00 crc kubenswrapper[4777]: I1124 17:01:00.385765 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:00 crc kubenswrapper[4777]: I1124 17:01:00.385781 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:00 crc kubenswrapper[4777]: I1124 17:01:00.426248 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 24 17:01:01 crc kubenswrapper[4777]: W1124 17:01:01.035734 4777 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 24 17:01:01 crc kubenswrapper[4777]: I1124 17:01:01.036139 4777 trace.go:236] Trace[948158659]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 17:00:51.033) (total time: 10002ms): Nov 24 17:01:01 crc kubenswrapper[4777]: Trace[948158659]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (17:01:01.035) Nov 24 17:01:01 crc kubenswrapper[4777]: Trace[948158659]: [10.002394863s] [10.002394863s] END Nov 24 17:01:01 crc kubenswrapper[4777]: E1124 17:01:01.036325 4777 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 24 17:01:01 crc kubenswrapper[4777]: I1124 17:01:01.166627 4777 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 24 17:01:01 crc kubenswrapper[4777]: I1124 17:01:01.375182 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:01:01 crc kubenswrapper[4777]: I1124 17:01:01.376367 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:01 crc kubenswrapper[4777]: I1124 17:01:01.376413 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:01 crc kubenswrapper[4777]: I1124 17:01:01.376430 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:01 crc kubenswrapper[4777]: I1124 17:01:01.392905 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 24 17:01:01 crc kubenswrapper[4777]: I1124 17:01:01.958063 4777 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 24 17:01:01 crc kubenswrapper[4777]: I1124 17:01:01.958152 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 24 17:01:02 crc kubenswrapper[4777]: W1124 17:01:02.061351 4777 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 24 17:01:02 crc kubenswrapper[4777]: I1124 17:01:02.061504 4777 trace.go:236] Trace[147873846]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 17:00:52.059) (total time: 10002ms): Nov 24 17:01:02 crc kubenswrapper[4777]: Trace[147873846]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (17:01:02.061) Nov 24 17:01:02 crc kubenswrapper[4777]: Trace[147873846]: [10.002163396s] [10.002163396s] END Nov 24 17:01:02 crc kubenswrapper[4777]: E1124 17:01:02.061536 4777 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 24 17:01:02 crc kubenswrapper[4777]: I1124 17:01:02.213806 4777 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 17:01:02 crc kubenswrapper[4777]: I1124 17:01:02.213896 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 17:01:02 crc kubenswrapper[4777]: I1124 17:01:02.219594 4777 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 17:01:02 crc kubenswrapper[4777]: I1124 17:01:02.219671 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 17:01:02 crc kubenswrapper[4777]: I1124 17:01:02.376939 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:01:02 crc kubenswrapper[4777]: I1124 17:01:02.377874 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:02 crc kubenswrapper[4777]: I1124 17:01:02.378180 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:02 crc kubenswrapper[4777]: I1124 17:01:02.378307 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:03 crc kubenswrapper[4777]: I1124 17:01:03.092502 4777 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]log ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]etcd ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/openshift.io-api-request-count-filter ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/openshift.io-startkubeinformers ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/generic-apiserver-start-informers ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/priority-and-fairness-config-consumer ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/priority-and-fairness-filter ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/start-apiextensions-informers ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/start-apiextensions-controllers ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/crd-informer-synced ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/start-system-namespaces-controller ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/start-cluster-authentication-info-controller ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/start-legacy-token-tracking-controller ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/start-service-ip-repair-controllers ok Nov 24 17:01:03 crc kubenswrapper[4777]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Nov 24 17:01:03 crc kubenswrapper[4777]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/priority-and-fairness-config-producer ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/bootstrap-controller ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/start-kube-aggregator-informers ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/apiservice-status-local-available-controller ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/apiservice-status-remote-available-controller ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/apiservice-registration-controller ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/apiservice-wait-for-first-sync ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/apiservice-discovery-controller ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/kube-apiserver-autoregistration ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]autoregister-completion ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/apiservice-openapi-controller ok Nov 24 17:01:03 crc kubenswrapper[4777]: [+]poststarthook/apiservice-openapiv3-controller ok Nov 24 17:01:03 crc kubenswrapper[4777]: livez check failed Nov 24 17:01:03 crc kubenswrapper[4777]: I1124 17:01:03.092569 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:01:03 crc kubenswrapper[4777]: E1124 17:01:03.403409 4777 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 17:01:07 crc kubenswrapper[4777]: E1124 17:01:07.205478 4777 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="7s" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.210044 4777 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 24 17:01:07 crc kubenswrapper[4777]: E1124 17:01:07.210049 4777 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.212049 4777 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.212766 4777 trace.go:236] Trace[2068990192]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 17:00:53.124) (total time: 14088ms): Nov 24 17:01:07 crc kubenswrapper[4777]: Trace[2068990192]: ---"Objects listed" error: 14087ms (17:01:07.212) Nov 24 17:01:07 crc kubenswrapper[4777]: Trace[2068990192]: [14.088004273s] [14.088004273s] END Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.212826 4777 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.324498 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.324649 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.325835 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.325893 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.325910 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.330744 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.399708 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.401297 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.404032 4777 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2" exitCode=255 Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.404125 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2"} Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.404217 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.404248 4777 scope.go:117] "RemoveContainer" containerID="6967922795b8929633b76e4ac97d832cae0f3f53600631b7cb52d16690f58e98" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.404416 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.405342 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.405387 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.405425 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.405435 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.405396 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.405525 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:07 crc kubenswrapper[4777]: I1124 17:01:07.406307 4777 scope.go:117] "RemoveContainer" containerID="673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2" Nov 24 17:01:07 crc kubenswrapper[4777]: E1124 17:01:07.406515 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 17:01:08 crc kubenswrapper[4777]: I1124 17:01:08.092511 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:01:08 crc kubenswrapper[4777]: I1124 17:01:08.117992 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:01:08 crc kubenswrapper[4777]: I1124 17:01:08.409502 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 24 17:01:08 crc kubenswrapper[4777]: I1124 17:01:08.412337 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:01:08 crc kubenswrapper[4777]: I1124 17:01:08.413511 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:08 crc kubenswrapper[4777]: I1124 17:01:08.413551 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:08 crc kubenswrapper[4777]: I1124 17:01:08.413564 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:08 crc kubenswrapper[4777]: I1124 17:01:08.414192 4777 scope.go:117] "RemoveContainer" containerID="673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2" Nov 24 17:01:08 crc kubenswrapper[4777]: E1124 17:01:08.414478 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 17:01:09 crc kubenswrapper[4777]: I1124 17:01:09.415076 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:01:09 crc kubenswrapper[4777]: I1124 17:01:09.416248 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:09 crc kubenswrapper[4777]: I1124 17:01:09.416284 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:09 crc kubenswrapper[4777]: I1124 17:01:09.416295 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:09 crc kubenswrapper[4777]: I1124 17:01:09.416988 4777 scope.go:117] "RemoveContainer" containerID="673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2" Nov 24 17:01:09 crc kubenswrapper[4777]: E1124 17:01:09.417169 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 17:01:10 crc kubenswrapper[4777]: I1124 17:01:10.301761 4777 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 24 17:01:10 crc kubenswrapper[4777]: I1124 17:01:10.789858 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:01:10 crc kubenswrapper[4777]: I1124 17:01:10.808551 4777 scope.go:117] "RemoveContainer" containerID="673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2" Nov 24 17:01:10 crc kubenswrapper[4777]: E1124 17:01:10.808772 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.149089 4777 apiserver.go:52] "Watching apiserver" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.153322 4777 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.153995 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-machine-config-operator/machine-config-daemon-vtgz9","openshift-multus/multus-additional-cni-plugins-xtmnl","openshift-multus/multus-mfvtr","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-node-identity/network-node-identity-vrzqb","openshift-ovn-kubernetes/ovnkube-node-djshf","openshift-dns/node-resolver-zd25z"] Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.154488 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.154513 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.155727 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.155876 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.156006 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.156876 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.157219 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.157293 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.157350 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.157406 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.157458 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.157460 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.157521 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.157980 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zd25z" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.158052 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.158677 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.158763 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.159431 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.162025 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.162400 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.162511 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.162576 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.162819 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.162885 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.162958 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.163017 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.163078 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.163238 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.163440 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.163823 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.164017 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.163855 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.164137 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.164205 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.164308 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.164353 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.164407 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.164145 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.164525 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.165215 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.165421 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.166544 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.166668 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.166857 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.167806 4777 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.169159 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.179151 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.193866 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.205785 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.232639 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.238653 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.238700 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.238724 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.238754 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.238778 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.238804 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.238826 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.238851 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.238874 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.238904 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.238926 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.238951 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.238996 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239021 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239046 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239067 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239087 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239110 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239130 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239153 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239177 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239200 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239223 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239248 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239249 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239271 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239300 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239334 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239319 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239361 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239386 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239416 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239443 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239473 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239502 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239486 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239526 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239552 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239486 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239575 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239606 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239630 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239650 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239673 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239684 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239695 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239724 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239746 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239767 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239816 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239839 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239861 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239884 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239904 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239922 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.239925 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240178 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240235 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240270 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240176 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240304 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240344 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240377 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240414 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240348 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240462 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240512 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240574 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240629 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240670 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240704 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240740 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240775 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240814 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240848 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240884 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240918 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240953 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241019 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241055 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241090 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241122 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241157 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241195 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241231 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241266 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241300 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241333 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241370 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240458 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241407 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241442 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241454 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241480 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241513 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241549 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241583 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241616 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241648 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241684 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241716 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241748 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241782 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241817 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241850 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241883 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241916 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.241952 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242005 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242041 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242073 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242109 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242143 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242175 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242210 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242223 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242246 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242281 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242373 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242401 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242406 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242459 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242533 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242591 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242643 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242699 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242710 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242755 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242751 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242803 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240917 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240951 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240993 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240946 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.242814 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.243006 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.243045 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.243078 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.243117 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.243146 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.243311 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.243403 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.243441 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.243676 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.243981 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.243948 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.244334 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.244382 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.244434 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.244443 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.244460 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.244604 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.244643 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.244670 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.244695 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.244702 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.244765 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.245063 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.245155 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.245356 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.245797 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.245175 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.245662 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.245881 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.246073 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.246136 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.246190 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.246235 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.246688 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.246739 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.246802 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.246864 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.246846 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.246881 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.247128 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.247195 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.247608 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.247678 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.247719 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.247761 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.247849 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.248224 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.248337 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.248371 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.248431 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.248675 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.248695 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.248719 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:01:11.74853707 +0000 UTC m=+29.907732159 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.248792 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.248802 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.248919 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.248987 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.249156 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.249246 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.249244 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.249279 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.240781 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.249502 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.249531 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.249890 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.249889 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.250300 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.250228 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.250357 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.250460 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.250354 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.250679 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.250733 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.250793 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.251056 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.251190 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.251535 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.251660 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.251715 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.251742 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.251739 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.251776 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.251807 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.251930 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252031 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252087 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252121 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252151 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252185 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252212 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252245 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252275 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252304 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252339 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252364 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252391 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252418 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252444 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252471 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252500 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252528 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252559 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252595 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252623 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252648 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252671 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252699 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252726 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252755 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252793 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252822 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252846 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252869 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252893 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252916 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252942 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252991 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253024 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253048 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253070 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253097 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253198 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253232 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253257 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253284 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253314 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253333 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253351 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253369 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253387 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253403 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253421 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253438 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253456 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253473 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253492 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253512 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253531 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253553 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253572 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253597 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253628 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253652 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253679 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253704 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253727 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253746 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253821 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253856 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-ovn\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253879 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253898 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0d953828-f624-4e22-890b-8a3a704547a7-mcd-auth-proxy-config\") pod \"machine-config-daemon-vtgz9\" (UID: \"0d953828-f624-4e22-890b-8a3a704547a7\") " pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253917 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-systemd-units\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253933 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253949 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-run-netns\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.253987 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/cce07485-e819-45d1-a91d-4bf1623a4324-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254015 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/54ae5e11-7c57-4c4f-a35a-4654f24ecdc3-hosts-file\") pod \"node-resolver-zd25z\" (UID: \"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\") " pod="openshift-dns/node-resolver-zd25z" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254036 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254058 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-cnibin\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254073 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-etc-kubernetes\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254092 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254110 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh8ww\" (UniqueName: \"kubernetes.io/projected/0d953828-f624-4e22-890b-8a3a704547a7-kube-api-access-zh8ww\") pod \"machine-config-daemon-vtgz9\" (UID: \"0d953828-f624-4e22-890b-8a3a704547a7\") " pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254126 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cce07485-e819-45d1-a91d-4bf1623a4324-os-release\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254142 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-openvswitch\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254160 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovnkube-config\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254176 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovnkube-script-lib\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254191 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-multus-conf-dir\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254208 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0d953828-f624-4e22-890b-8a3a704547a7-rootfs\") pod \"machine-config-daemon-vtgz9\" (UID: \"0d953828-f624-4e22-890b-8a3a704547a7\") " pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254224 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cce07485-e819-45d1-a91d-4bf1623a4324-cnibin\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254240 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-run-netns\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254260 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovn-node-metrics-cert\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.254986 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.255912 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252468 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.257884 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252543 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252690 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.252952 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.254792 4777 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.258091 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.255858 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.255893 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.256339 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.256494 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.256499 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.256872 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.256922 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.257097 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.257224 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.257276 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.257319 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.257807 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.257831 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.258220 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:11.758169653 +0000 UTC m=+29.917364732 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.258259 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.258302 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.258842 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.259012 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.259221 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.259286 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.259563 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.259841 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.259945 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cce07485-e819-45d1-a91d-4bf1623a4324-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.259942 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.260004 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.260247 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.260181 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.261656 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.261431 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.263014 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.263185 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.263294 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.263516 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.263862 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.264034 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.264107 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.264367 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.264551 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.264431 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-systemd\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.264810 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.264847 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-env-overrides\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.264947 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.265042 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.265053 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.265085 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cce07485-e819-45d1-a91d-4bf1623a4324-cni-binary-copy\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.265188 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-system-cni-dir\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.265214 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a5500f09-0e35-442c-87ae-8f280cd2edd1-cni-binary-copy\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.265238 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-run-k8s-cni-cncf-io\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.265306 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-var-lib-kubelet\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.265464 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.265461 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.265517 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-slash\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.265578 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.265744 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.265997 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.266043 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.266061 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.266334 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.266359 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.266423 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.266476 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.266501 4777 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.266697 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.266757 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.266883 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.267038 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.267239 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:11.767163875 +0000 UTC m=+29.926358944 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.267466 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.267584 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.267724 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.267857 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.268745 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7knbn\" (UniqueName: \"kubernetes.io/projected/cce07485-e819-45d1-a91d-4bf1623a4324-kube-api-access-7knbn\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.269027 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.269583 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-log-socket\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.269659 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.271059 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.270539 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.271089 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.271138 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.271162 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.271242 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.271469 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.271508 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.269724 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.271739 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cce07485-e819-45d1-a91d-4bf1623a4324-system-cni-dir\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.271818 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-kubelet\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.271882 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhmkp\" (UniqueName: \"kubernetes.io/projected/a5500f09-0e35-442c-87ae-8f280cd2edd1-kube-api-access-jhmkp\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.271926 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0d953828-f624-4e22-890b-8a3a704547a7-proxy-tls\") pod \"machine-config-daemon-vtgz9\" (UID: \"0d953828-f624-4e22-890b-8a3a704547a7\") " pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.271952 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-multus-cni-dir\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.272057 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-var-lib-cni-bin\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.272098 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-hostroot\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.272177 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-var-lib-openvswitch\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.272049 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.272142 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.272222 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-etc-openvswitch\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.272247 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk9ks\" (UniqueName: \"kubernetes.io/projected/9084c33d-bd72-470c-9f45-3b0b0f29aa19-kube-api-access-kk9ks\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.272280 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.272383 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-node-log\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.272407 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-os-release\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.272436 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.272450 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-var-lib-cni-multus\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.272502 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.272732 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.272937 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.273057 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.273064 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.273096 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.273610 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.274141 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.275393 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-run-ovn-kubernetes\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.275563 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-cni-bin\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.275598 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-cni-netd\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.275673 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-run-multus-certs\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.275745 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llg6t\" (UniqueName: \"kubernetes.io/projected/54ae5e11-7c57-4c4f-a35a-4654f24ecdc3-kube-api-access-llg6t\") pod \"node-resolver-zd25z\" (UID: \"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\") " pod="openshift-dns/node-resolver-zd25z" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.275778 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-multus-socket-dir-parent\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.275196 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.274617 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.274660 4777 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.275857 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/a5500f09-0e35-442c-87ae-8f280cd2edd1-multus-daemon-config\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.276190 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.277418 4777 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.277448 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.277508 4777 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.277529 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.277547 4777 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.277683 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.277740 4777 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.277993 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278019 4777 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278037 4777 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278052 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278068 4777 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278085 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278101 4777 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278115 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278128 4777 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278144 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278158 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278171 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278185 4777 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278201 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278217 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278231 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278246 4777 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278260 4777 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278275 4777 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278289 4777 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278306 4777 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278320 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278335 4777 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278350 4777 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278364 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278378 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278391 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278404 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278420 4777 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278433 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278447 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278461 4777 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278475 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278491 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278504 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278519 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278532 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278546 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278559 4777 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278572 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278585 4777 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278598 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278610 4777 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278622 4777 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278635 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278649 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278662 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278675 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278691 4777 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278704 4777 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278718 4777 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278731 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278768 4777 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278783 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278800 4777 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278818 4777 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278835 4777 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278864 4777 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278883 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278898 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278910 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278922 4777 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278934 4777 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278946 4777 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.278960 4777 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279367 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279381 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279395 4777 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279410 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279424 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279436 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279449 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279462 4777 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279475 4777 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279488 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279503 4777 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279517 4777 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279530 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279543 4777 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279556 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279569 4777 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279581 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279594 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279606 4777 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279623 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279651 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279664 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279676 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279690 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279704 4777 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279718 4777 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279733 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279747 4777 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279760 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279772 4777 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279785 4777 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279798 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279811 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279826 4777 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279838 4777 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279851 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279865 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279877 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279890 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279903 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279916 4777 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279928 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279941 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279953 4777 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.279998 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280011 4777 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280022 4777 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280034 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280047 4777 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280059 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280071 4777 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280084 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.269864 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.269918 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.270136 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.270166 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.270737 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.270791 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.270876 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280124 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280248 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280265 4777 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280299 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280314 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280329 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280344 4777 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280359 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280374 4777 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280387 4777 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280400 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280413 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280427 4777 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280440 4777 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280452 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280467 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280510 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280528 4777 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280543 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280557 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280570 4777 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280583 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280596 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280609 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280622 4777 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280636 4777 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280674 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280702 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280716 4777 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280729 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280755 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280769 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280793 4777 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280805 4777 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280819 4777 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280832 4777 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.280881 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.281197 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.284416 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.285795 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.290084 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.286687 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.290751 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.291477 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.292089 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.292183 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.292847 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.293404 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.293536 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.293588 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.293621 4777 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.293325 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.293686 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.294447 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.294459 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.294496 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.293757 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:11.793690908 +0000 UTC m=+29.952885967 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.295733 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.298400 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.298564 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.298640 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.299102 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.299277 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.299559 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.299595 4777 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.299772 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:11.799653586 +0000 UTC m=+29.958848645 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.301358 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.301707 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.303590 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.304104 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.305816 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.304153 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.298602 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.305189 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.305270 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.308814 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.309891 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.310872 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.313125 4777 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.313463 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.314696 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.315891 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.321063 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.321807 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.323315 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.324568 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.326668 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.327392 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.327798 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.330064 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.332998 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.333847 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.334270 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.335701 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.336759 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.338427 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.340114 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.341000 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.340922 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.342095 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.343431 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.343556 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.345462 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.346230 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.346811 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.347919 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.348677 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.349957 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.350920 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.351619 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.361437 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.371777 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382553 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-var-lib-openvswitch\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382586 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-etc-openvswitch\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382605 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk9ks\" (UniqueName: \"kubernetes.io/projected/9084c33d-bd72-470c-9f45-3b0b0f29aa19-kube-api-access-kk9ks\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382630 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-node-log\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382649 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-os-release\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382665 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-var-lib-cni-multus\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382681 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-cni-netd\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382701 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-run-multus-certs\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382718 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382733 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-run-ovn-kubernetes\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382753 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-cni-bin\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382770 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llg6t\" (UniqueName: \"kubernetes.io/projected/54ae5e11-7c57-4c4f-a35a-4654f24ecdc3-kube-api-access-llg6t\") pod \"node-resolver-zd25z\" (UID: \"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\") " pod="openshift-dns/node-resolver-zd25z" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382785 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-multus-socket-dir-parent\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382806 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/a5500f09-0e35-442c-87ae-8f280cd2edd1-multus-daemon-config\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382821 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-ovn\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382839 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0d953828-f624-4e22-890b-8a3a704547a7-mcd-auth-proxy-config\") pod \"machine-config-daemon-vtgz9\" (UID: \"0d953828-f624-4e22-890b-8a3a704547a7\") " pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382860 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-systemd-units\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382880 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382897 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-run-netns\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382930 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-cnibin\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382945 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-etc-kubernetes\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.382987 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/cce07485-e819-45d1-a91d-4bf1623a4324-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383008 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/54ae5e11-7c57-4c4f-a35a-4654f24ecdc3-hosts-file\") pod \"node-resolver-zd25z\" (UID: \"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\") " pod="openshift-dns/node-resolver-zd25z" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383027 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovnkube-script-lib\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383073 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-multus-conf-dir\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383102 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh8ww\" (UniqueName: \"kubernetes.io/projected/0d953828-f624-4e22-890b-8a3a704547a7-kube-api-access-zh8ww\") pod \"machine-config-daemon-vtgz9\" (UID: \"0d953828-f624-4e22-890b-8a3a704547a7\") " pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383120 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cce07485-e819-45d1-a91d-4bf1623a4324-os-release\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383138 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-openvswitch\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383174 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovnkube-config\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383197 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0d953828-f624-4e22-890b-8a3a704547a7-rootfs\") pod \"machine-config-daemon-vtgz9\" (UID: \"0d953828-f624-4e22-890b-8a3a704547a7\") " pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383216 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cce07485-e819-45d1-a91d-4bf1623a4324-cnibin\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383273 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-run-netns\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383260 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-var-lib-cni-multus\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383297 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovn-node-metrics-cert\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383349 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-etc-kubernetes\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383347 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383413 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-run-ovn-kubernetes\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383454 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-node-log\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383495 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-etc-openvswitch\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383511 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-multus-socket-dir-parent\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383754 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cce07485-e819-45d1-a91d-4bf1623a4324-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384640 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cce07485-e819-45d1-a91d-4bf1623a4324-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384672 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-systemd\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383803 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-cni-netd\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383864 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384709 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-env-overrides\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383929 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-run-netns\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384116 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/54ae5e11-7c57-4c4f-a35a-4654f24ecdc3-hosts-file\") pod \"node-resolver-zd25z\" (UID: \"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\") " pod="openshift-dns/node-resolver-zd25z" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384745 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-run-k8s-cni-cncf-io\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384123 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-ovn\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384777 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-var-lib-openvswitch\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384783 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-var-lib-kubelet\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384178 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cce07485-e819-45d1-a91d-4bf1623a4324-os-release\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384200 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cce07485-e819-45d1-a91d-4bf1623a4324-cnibin\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384217 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-run-netns\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384239 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0d953828-f624-4e22-890b-8a3a704547a7-rootfs\") pod \"machine-config-daemon-vtgz9\" (UID: \"0d953828-f624-4e22-890b-8a3a704547a7\") " pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383829 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-run-multus-certs\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384849 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cce07485-e819-45d1-a91d-4bf1623a4324-cni-binary-copy\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384780 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-os-release\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384877 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-run-k8s-cni-cncf-io\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384142 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-cnibin\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384191 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-systemd-units\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384893 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-system-cni-dir\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384907 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-var-lib-kubelet\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384045 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-openvswitch\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.384142 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-multus-conf-dir\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.383843 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-cni-bin\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385018 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-systemd\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385028 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a5500f09-0e35-442c-87ae-8f280cd2edd1-cni-binary-copy\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385060 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-system-cni-dir\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385052 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/a5500f09-0e35-442c-87ae-8f280cd2edd1-multus-daemon-config\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385118 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-slash\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385142 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-slash\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385182 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385426 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7knbn\" (UniqueName: \"kubernetes.io/projected/cce07485-e819-45d1-a91d-4bf1623a4324-kube-api-access-7knbn\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385456 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-log-socket\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385488 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cce07485-e819-45d1-a91d-4bf1623a4324-system-cni-dir\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385511 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-kubelet\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385530 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhmkp\" (UniqueName: \"kubernetes.io/projected/a5500f09-0e35-442c-87ae-8f280cd2edd1-kube-api-access-jhmkp\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385550 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0d953828-f624-4e22-890b-8a3a704547a7-proxy-tls\") pod \"machine-config-daemon-vtgz9\" (UID: \"0d953828-f624-4e22-890b-8a3a704547a7\") " pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385567 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-multus-cni-dir\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385584 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-var-lib-cni-bin\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385625 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-hostroot\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385689 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-log-socket\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385695 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-kubelet\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385426 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.385896 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/cce07485-e819-45d1-a91d-4bf1623a4324-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.386334 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cce07485-e819-45d1-a91d-4bf1623a4324-cni-binary-copy\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.386421 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-env-overrides\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.386778 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-multus-cni-dir\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.386803 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovnkube-script-lib\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.386854 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cce07485-e819-45d1-a91d-4bf1623a4324-system-cni-dir\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.386882 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-host-var-lib-cni-bin\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.386916 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/a5500f09-0e35-442c-87ae-8f280cd2edd1-hostroot\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.387037 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.387062 4777 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.387080 4777 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.387097 4777 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.387110 4777 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.387124 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.387139 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393200 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393227 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393244 4777 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393259 4777 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393274 4777 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393288 4777 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393301 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393317 4777 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393337 4777 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393354 4777 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393369 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393384 4777 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393397 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393410 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393424 4777 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393437 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393455 4777 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393470 4777 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393484 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393503 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393517 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393530 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393544 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393558 4777 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393572 4777 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393589 4777 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.393608 4777 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.388173 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a5500f09-0e35-442c-87ae-8f280cd2edd1-cni-binary-copy\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.387284 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovnkube-config\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.387457 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0d953828-f624-4e22-890b-8a3a704547a7-mcd-auth-proxy-config\") pod \"machine-config-daemon-vtgz9\" (UID: \"0d953828-f624-4e22-890b-8a3a704547a7\") " pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.396370 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovn-node-metrics-cert\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.396491 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-h8twn"] Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.396524 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0d953828-f624-4e22-890b-8a3a704547a7-proxy-tls\") pod \"machine-config-daemon-vtgz9\" (UID: \"0d953828-f624-4e22-890b-8a3a704547a7\") " pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.397045 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-h8twn" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.400463 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.400573 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.400804 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.401009 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.402013 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llg6t\" (UniqueName: \"kubernetes.io/projected/54ae5e11-7c57-4c4f-a35a-4654f24ecdc3-kube-api-access-llg6t\") pod \"node-resolver-zd25z\" (UID: \"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\") " pod="openshift-dns/node-resolver-zd25z" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.404526 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk9ks\" (UniqueName: \"kubernetes.io/projected/9084c33d-bd72-470c-9f45-3b0b0f29aa19-kube-api-access-kk9ks\") pod \"ovnkube-node-djshf\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.404520 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7knbn\" (UniqueName: \"kubernetes.io/projected/cce07485-e819-45d1-a91d-4bf1623a4324-kube-api-access-7knbn\") pod \"multus-additional-cni-plugins-xtmnl\" (UID: \"cce07485-e819-45d1-a91d-4bf1623a4324\") " pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.405588 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh8ww\" (UniqueName: \"kubernetes.io/projected/0d953828-f624-4e22-890b-8a3a704547a7-kube-api-access-zh8ww\") pod \"machine-config-daemon-vtgz9\" (UID: \"0d953828-f624-4e22-890b-8a3a704547a7\") " pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.408150 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.411447 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhmkp\" (UniqueName: \"kubernetes.io/projected/a5500f09-0e35-442c-87ae-8f280cd2edd1-kube-api-access-jhmkp\") pod \"multus-mfvtr\" (UID: \"a5500f09-0e35-442c-87ae-8f280cd2edd1\") " pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.420728 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.421834 4777 scope.go:117] "RemoveContainer" containerID="673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.422152 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.431842 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.451798 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.463860 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.472385 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.477045 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.480553 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.486345 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" Nov 24 17:01:11 crc kubenswrapper[4777]: W1124 17:01:11.491246 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-4be369a7a2bcd55fe2b5468cac100d2b2dac022b6e50cddbb28c4aff44b15275 WatchSource:0}: Error finding container 4be369a7a2bcd55fe2b5468cac100d2b2dac022b6e50cddbb28c4aff44b15275: Status 404 returned error can't find the container with id 4be369a7a2bcd55fe2b5468cac100d2b2dac022b6e50cddbb28c4aff44b15275 Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.492493 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.494724 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-mfvtr" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.495074 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d2d91d83-93ba-4997-8e57-bbefc542c8f4-serviceca\") pod \"node-ca-h8twn\" (UID: \"d2d91d83-93ba-4997-8e57-bbefc542c8f4\") " pod="openshift-image-registry/node-ca-h8twn" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.495144 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s6jc\" (UniqueName: \"kubernetes.io/projected/d2d91d83-93ba-4997-8e57-bbefc542c8f4-kube-api-access-9s6jc\") pod \"node-ca-h8twn\" (UID: \"d2d91d83-93ba-4997-8e57-bbefc542c8f4\") " pod="openshift-image-registry/node-ca-h8twn" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.495177 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2d91d83-93ba-4997-8e57-bbefc542c8f4-host\") pod \"node-ca-h8twn\" (UID: \"d2d91d83-93ba-4997-8e57-bbefc542c8f4\") " pod="openshift-image-registry/node-ca-h8twn" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.504695 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.511230 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: W1124 17:01:11.512006 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcce07485_e819_45d1_a91d_4bf1623a4324.slice/crio-2c00892cd9afdd5520955e29343f6b6d9da641980976e3386f0a6c8c403cba29 WatchSource:0}: Error finding container 2c00892cd9afdd5520955e29343f6b6d9da641980976e3386f0a6c8c403cba29: Status 404 returned error can't find the container with id 2c00892cd9afdd5520955e29343f6b6d9da641980976e3386f0a6c8c403cba29 Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.514747 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.524518 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zd25z" Nov 24 17:01:11 crc kubenswrapper[4777]: W1124 17:01:11.526335 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5500f09_0e35_442c_87ae_8f280cd2edd1.slice/crio-311b885711b03953f5115269866152a11a4744dfae8ab89f3dba76a2a3c1d452 WatchSource:0}: Error finding container 311b885711b03953f5115269866152a11a4744dfae8ab89f3dba76a2a3c1d452: Status 404 returned error can't find the container with id 311b885711b03953f5115269866152a11a4744dfae8ab89f3dba76a2a3c1d452 Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.530360 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.534958 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:11 crc kubenswrapper[4777]: W1124 17:01:11.540960 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d953828_f624_4e22_890b_8a3a704547a7.slice/crio-e2c964dfd66c0a245de0a88c5d79458aa9e521748f72dc30ac2890b7190f6826 WatchSource:0}: Error finding container e2c964dfd66c0a245de0a88c5d79458aa9e521748f72dc30ac2890b7190f6826: Status 404 returned error can't find the container with id e2c964dfd66c0a245de0a88c5d79458aa9e521748f72dc30ac2890b7190f6826 Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.548586 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.561692 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: W1124 17:01:11.565848 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-79746c9f0abab99fb4c79722f4a216c35bb6d63e56cfb6a2ecf096429c399df3 WatchSource:0}: Error finding container 79746c9f0abab99fb4c79722f4a216c35bb6d63e56cfb6a2ecf096429c399df3: Status 404 returned error can't find the container with id 79746c9f0abab99fb4c79722f4a216c35bb6d63e56cfb6a2ecf096429c399df3 Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.573186 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.587775 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 17:01:11 crc kubenswrapper[4777]: W1124 17:01:11.591088 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9084c33d_bd72_470c_9f45_3b0b0f29aa19.slice/crio-93a349ce411ab44f6eb3c91d06db3e86b9dc6d7413f6af0b8575c8c9c0253ffd WatchSource:0}: Error finding container 93a349ce411ab44f6eb3c91d06db3e86b9dc6d7413f6af0b8575c8c9c0253ffd: Status 404 returned error can't find the container with id 93a349ce411ab44f6eb3c91d06db3e86b9dc6d7413f6af0b8575c8c9c0253ffd Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.595734 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s6jc\" (UniqueName: \"kubernetes.io/projected/d2d91d83-93ba-4997-8e57-bbefc542c8f4-kube-api-access-9s6jc\") pod \"node-ca-h8twn\" (UID: \"d2d91d83-93ba-4997-8e57-bbefc542c8f4\") " pod="openshift-image-registry/node-ca-h8twn" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.595807 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2d91d83-93ba-4997-8e57-bbefc542c8f4-host\") pod \"node-ca-h8twn\" (UID: \"d2d91d83-93ba-4997-8e57-bbefc542c8f4\") " pod="openshift-image-registry/node-ca-h8twn" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.595856 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d2d91d83-93ba-4997-8e57-bbefc542c8f4-serviceca\") pod \"node-ca-h8twn\" (UID: \"d2d91d83-93ba-4997-8e57-bbefc542c8f4\") " pod="openshift-image-registry/node-ca-h8twn" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.595897 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2d91d83-93ba-4997-8e57-bbefc542c8f4-host\") pod \"node-ca-h8twn\" (UID: \"d2d91d83-93ba-4997-8e57-bbefc542c8f4\") " pod="openshift-image-registry/node-ca-h8twn" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.597101 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d2d91d83-93ba-4997-8e57-bbefc542c8f4-serviceca\") pod \"node-ca-h8twn\" (UID: \"d2d91d83-93ba-4997-8e57-bbefc542c8f4\") " pod="openshift-image-registry/node-ca-h8twn" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.622049 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s6jc\" (UniqueName: \"kubernetes.io/projected/d2d91d83-93ba-4997-8e57-bbefc542c8f4-kube-api-access-9s6jc\") pod \"node-ca-h8twn\" (UID: \"d2d91d83-93ba-4997-8e57-bbefc542c8f4\") " pod="openshift-image-registry/node-ca-h8twn" Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.724145 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-h8twn" Nov 24 17:01:11 crc kubenswrapper[4777]: W1124 17:01:11.796006 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2d91d83_93ba_4997_8e57_bbefc542c8f4.slice/crio-d15bcefe5720e62e3e87252405373a4da47bbc44c3a72f6c868f7f7139d4e7ba WatchSource:0}: Error finding container d15bcefe5720e62e3e87252405373a4da47bbc44c3a72f6c868f7f7139d4e7ba: Status 404 returned error can't find the container with id d15bcefe5720e62e3e87252405373a4da47bbc44c3a72f6c868f7f7139d4e7ba Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.797824 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.797950 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.798093 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:01:12.798056981 +0000 UTC m=+30.957252130 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.798139 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.798169 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.798174 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.798184 4777 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.798707 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.798783 4777 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.798791 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:12.798777534 +0000 UTC m=+30.957972583 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.798229 4777 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.798845 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:12.798827485 +0000 UTC m=+30.958022544 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.798871 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:12.798861266 +0000 UTC m=+30.958056325 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: I1124 17:01:11.903406 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.903595 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.903629 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.903645 4777 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:11 crc kubenswrapper[4777]: E1124 17:01:11.903701 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:12.903685799 +0000 UTC m=+31.062880858 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.244883 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:12 crc kubenswrapper[4777]: E1124 17:01:12.245163 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.245457 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:12 crc kubenswrapper[4777]: E1124 17:01:12.245726 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.427085 4777 generic.go:334] "Generic (PLEG): container finished" podID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerID="94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc" exitCode=0 Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.427202 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerDied","Data":"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.427290 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerStarted","Data":"93a349ce411ab44f6eb3c91d06db3e86b9dc6d7413f6af0b8575c8c9c0253ffd"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.431185 4777 generic.go:334] "Generic (PLEG): container finished" podID="cce07485-e819-45d1-a91d-4bf1623a4324" containerID="a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862" exitCode=0 Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.431250 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" event={"ID":"cce07485-e819-45d1-a91d-4bf1623a4324","Type":"ContainerDied","Data":"a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.431363 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" event={"ID":"cce07485-e819-45d1-a91d-4bf1623a4324","Type":"ContainerStarted","Data":"2c00892cd9afdd5520955e29343f6b6d9da641980976e3386f0a6c8c403cba29"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.433234 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zd25z" event={"ID":"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3","Type":"ContainerStarted","Data":"168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.433276 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zd25z" event={"ID":"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3","Type":"ContainerStarted","Data":"3a9cdcebdd5d1db3f1b8085de7c2e066a5f595d843eeb122def4b5a28b7999fc"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.435233 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"79746c9f0abab99fb4c79722f4a216c35bb6d63e56cfb6a2ecf096429c399df3"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.437615 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-h8twn" event={"ID":"d2d91d83-93ba-4997-8e57-bbefc542c8f4","Type":"ContainerStarted","Data":"5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.437673 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-h8twn" event={"ID":"d2d91d83-93ba-4997-8e57-bbefc542c8f4","Type":"ContainerStarted","Data":"d15bcefe5720e62e3e87252405373a4da47bbc44c3a72f6c868f7f7139d4e7ba"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.442650 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mfvtr" event={"ID":"a5500f09-0e35-442c-87ae-8f280cd2edd1","Type":"ContainerStarted","Data":"dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.443015 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mfvtr" event={"ID":"a5500f09-0e35-442c-87ae-8f280cd2edd1","Type":"ContainerStarted","Data":"311b885711b03953f5115269866152a11a4744dfae8ab89f3dba76a2a3c1d452"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.446536 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.446606 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"4be369a7a2bcd55fe2b5468cac100d2b2dac022b6e50cddbb28c4aff44b15275"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.450765 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.450813 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.450828 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"e2c964dfd66c0a245de0a88c5d79458aa9e521748f72dc30ac2890b7190f6826"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.453675 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.453751 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.453769 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a01e707b083ff7c3911c6776c7209de14bc13d67bfa2114ae8d90b112992a094"} Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.459310 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.476275 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.498516 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.519523 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.539018 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.568103 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.593273 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.625495 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.658393 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.678318 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.693711 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.705111 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.727356 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.746234 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.760120 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.773436 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.787890 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.801149 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.814860 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.817012 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.817269 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:12 crc kubenswrapper[4777]: E1124 17:01:12.817381 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:01:14.817345107 +0000 UTC m=+32.976540166 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:01:12 crc kubenswrapper[4777]: E1124 17:01:12.817445 4777 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.817474 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:12 crc kubenswrapper[4777]: E1124 17:01:12.817541 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:14.817510912 +0000 UTC m=+32.976705991 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.817591 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:12 crc kubenswrapper[4777]: E1124 17:01:12.817610 4777 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:01:12 crc kubenswrapper[4777]: E1124 17:01:12.817705 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:14.817677678 +0000 UTC m=+32.976872737 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:01:12 crc kubenswrapper[4777]: E1124 17:01:12.817734 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:01:12 crc kubenswrapper[4777]: E1124 17:01:12.817757 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:01:12 crc kubenswrapper[4777]: E1124 17:01:12.817777 4777 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:12 crc kubenswrapper[4777]: E1124 17:01:12.817825 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:14.817810742 +0000 UTC m=+32.977005831 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.841392 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.852823 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.870998 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.886289 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.908661 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.919282 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:12 crc kubenswrapper[4777]: E1124 17:01:12.919546 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:01:12 crc kubenswrapper[4777]: E1124 17:01:12.919597 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:01:12 crc kubenswrapper[4777]: E1124 17:01:12.919615 4777 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:12 crc kubenswrapper[4777]: E1124 17:01:12.919691 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:14.919669611 +0000 UTC m=+33.078864660 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.922820 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:12 crc kubenswrapper[4777]: I1124 17:01:12.934213 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:12Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.161759 4777 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.244598 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:13 crc kubenswrapper[4777]: E1124 17:01:13.245283 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.252374 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.253730 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.256601 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.258208 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.261200 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.262807 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.264565 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.265603 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.266556 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.267291 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.267989 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.268911 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.269801 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.271090 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.271785 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.273199 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.274010 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.275280 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.276109 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.276816 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.278073 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.278891 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.288780 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.302227 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.311963 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.326193 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.340242 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.355802 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.370485 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.385411 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.405951 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.422845 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.444294 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.458861 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerStarted","Data":"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807"} Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.458909 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerStarted","Data":"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b"} Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.458920 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerStarted","Data":"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366"} Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.460734 4777 generic.go:334] "Generic (PLEG): container finished" podID="cce07485-e819-45d1-a91d-4bf1623a4324" containerID="f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5" exitCode=0 Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.460784 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" event={"ID":"cce07485-e819-45d1-a91d-4bf1623a4324","Type":"ContainerDied","Data":"f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5"} Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.464952 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.493442 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.512123 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.536905 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.548337 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.563611 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.577135 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.594800 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.608250 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.619349 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.636920 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.652852 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.665945 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:13 crc kubenswrapper[4777]: I1124 17:01:13.683280 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.210792 4777 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.213189 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.213244 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.213264 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.213415 4777 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.223272 4777 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.223736 4777 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.225318 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.225369 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.225381 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.225399 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.225411 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:14Z","lastTransitionTime":"2025-11-24T17:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.243446 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.244837 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.244919 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.244988 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.245161 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.248500 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.248550 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.248565 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.248588 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.248604 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:14Z","lastTransitionTime":"2025-11-24T17:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.263614 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.268329 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.268419 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.268437 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.268466 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.268487 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:14Z","lastTransitionTime":"2025-11-24T17:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.288727 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.293996 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.294044 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.294064 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.294092 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.294113 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:14Z","lastTransitionTime":"2025-11-24T17:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.306651 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.310921 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.311015 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.311035 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.311061 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.311080 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:14Z","lastTransitionTime":"2025-11-24T17:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.333737 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.333865 4777 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.336033 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.336087 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.336108 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.336133 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.336152 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:14Z","lastTransitionTime":"2025-11-24T17:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.439356 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.439395 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.439404 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.439420 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.439430 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:14Z","lastTransitionTime":"2025-11-24T17:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.469355 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerStarted","Data":"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a"} Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.469435 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerStarted","Data":"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860"} Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.469451 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerStarted","Data":"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab"} Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.471402 4777 generic.go:334] "Generic (PLEG): container finished" podID="cce07485-e819-45d1-a91d-4bf1623a4324" containerID="623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e" exitCode=0 Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.471450 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" event={"ID":"cce07485-e819-45d1-a91d-4bf1623a4324","Type":"ContainerDied","Data":"623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e"} Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.490139 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.509771 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.525120 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.543038 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.543093 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.543110 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.543137 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.543154 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:14Z","lastTransitionTime":"2025-11-24T17:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.546777 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.572436 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.591645 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.614024 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.631048 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.647022 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.647093 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.647110 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.647136 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.647153 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:14Z","lastTransitionTime":"2025-11-24T17:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.655247 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.668470 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.686329 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.703357 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.719498 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:14Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.750502 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.750555 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.750572 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.750596 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.750610 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:14Z","lastTransitionTime":"2025-11-24T17:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.845010 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.845140 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.845183 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.845216 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.845276 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:01:18.845229553 +0000 UTC m=+37.004424642 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.845348 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.845431 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.845446 4777 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.845510 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:18.845494501 +0000 UTC m=+37.004689560 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.845528 4777 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.845565 4777 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.845669 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:18.845634736 +0000 UTC m=+37.004829825 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.845799 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:18.845689507 +0000 UTC m=+37.004884656 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.858294 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.858334 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.858346 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.858365 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.858379 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:14Z","lastTransitionTime":"2025-11-24T17:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.946435 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.946796 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.946872 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.946905 4777 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:14 crc kubenswrapper[4777]: E1124 17:01:14.947069 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:18.947030691 +0000 UTC m=+37.106225900 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.961440 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.961494 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.961514 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.961540 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:14 crc kubenswrapper[4777]: I1124 17:01:14.961559 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:14Z","lastTransitionTime":"2025-11-24T17:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.066312 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.066374 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.066385 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.066406 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.066422 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:15Z","lastTransitionTime":"2025-11-24T17:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.170021 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.170071 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.170081 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.170100 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.170114 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:15Z","lastTransitionTime":"2025-11-24T17:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.245178 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:15 crc kubenswrapper[4777]: E1124 17:01:15.245443 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.273371 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.273459 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.273479 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.273507 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.273525 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:15Z","lastTransitionTime":"2025-11-24T17:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.376861 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.376956 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.377015 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.377048 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.377075 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:15Z","lastTransitionTime":"2025-11-24T17:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.476709 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c"} Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.479425 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.479488 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.479505 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.479527 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.479546 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:15Z","lastTransitionTime":"2025-11-24T17:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.481087 4777 generic.go:334] "Generic (PLEG): container finished" podID="cce07485-e819-45d1-a91d-4bf1623a4324" containerID="8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412" exitCode=0 Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.481164 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" event={"ID":"cce07485-e819-45d1-a91d-4bf1623a4324","Type":"ContainerDied","Data":"8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412"} Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.502851 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.529132 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.550944 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.569381 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.603293 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.606056 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.606105 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.606127 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.606156 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.606178 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:15Z","lastTransitionTime":"2025-11-24T17:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.633012 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.655720 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.667463 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.677786 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.688784 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.698437 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.708006 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.708132 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.708193 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.708262 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.708318 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:15Z","lastTransitionTime":"2025-11-24T17:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.713879 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.723046 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.734808 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.750594 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.764615 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.777198 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.789427 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.810830 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.810890 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.810906 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.810934 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.810953 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:15Z","lastTransitionTime":"2025-11-24T17:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.819376 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.840083 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.859591 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.887709 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.911012 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.914203 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.914284 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.914301 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.914323 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.914338 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:15Z","lastTransitionTime":"2025-11-24T17:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.932608 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.949444 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:15 crc kubenswrapper[4777]: I1124 17:01:15.969765 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:15Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.016848 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.016912 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.016937 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.016961 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.017007 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:16Z","lastTransitionTime":"2025-11-24T17:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.119133 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.119195 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.119211 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.119278 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.119302 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:16Z","lastTransitionTime":"2025-11-24T17:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.221768 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.221804 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.221813 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.221827 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.221838 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:16Z","lastTransitionTime":"2025-11-24T17:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.244196 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.244220 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:16 crc kubenswrapper[4777]: E1124 17:01:16.244328 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:16 crc kubenswrapper[4777]: E1124 17:01:16.244511 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.324472 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.324553 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.324572 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.324598 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.324616 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:16Z","lastTransitionTime":"2025-11-24T17:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.427828 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.427906 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.427931 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.427962 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.428017 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:16Z","lastTransitionTime":"2025-11-24T17:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.491178 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerStarted","Data":"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936"} Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.496012 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" event={"ID":"cce07485-e819-45d1-a91d-4bf1623a4324","Type":"ContainerDied","Data":"30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1"} Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.496024 4777 generic.go:334] "Generic (PLEG): container finished" podID="cce07485-e819-45d1-a91d-4bf1623a4324" containerID="30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1" exitCode=0 Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.517500 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.531999 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.532045 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.532064 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.532087 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.532106 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:16Z","lastTransitionTime":"2025-11-24T17:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.538302 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.558434 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.581909 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.603386 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.633928 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.636104 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.636139 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.636154 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.636174 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.636188 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:16Z","lastTransitionTime":"2025-11-24T17:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.656209 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.677572 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.699942 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.726332 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.738755 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.738791 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.738802 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.738820 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.738833 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:16Z","lastTransitionTime":"2025-11-24T17:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.744498 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.759701 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.781714 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.841742 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.841788 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.841798 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.841813 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.841823 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:16Z","lastTransitionTime":"2025-11-24T17:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.945257 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.945300 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.945310 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.945324 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:16 crc kubenswrapper[4777]: I1124 17:01:16.945335 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:16Z","lastTransitionTime":"2025-11-24T17:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.048477 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.048549 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.048563 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.048588 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.048603 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:17Z","lastTransitionTime":"2025-11-24T17:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.151762 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.151821 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.151834 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.151855 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.151870 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:17Z","lastTransitionTime":"2025-11-24T17:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.245205 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:17 crc kubenswrapper[4777]: E1124 17:01:17.245440 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.254557 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.254591 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.254600 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.254611 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.254622 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:17Z","lastTransitionTime":"2025-11-24T17:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.358067 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.358131 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.358149 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.358175 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.358194 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:17Z","lastTransitionTime":"2025-11-24T17:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.462280 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.462341 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.462362 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.462389 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.462408 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:17Z","lastTransitionTime":"2025-11-24T17:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.505204 4777 generic.go:334] "Generic (PLEG): container finished" podID="cce07485-e819-45d1-a91d-4bf1623a4324" containerID="862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9" exitCode=0 Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.505274 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" event={"ID":"cce07485-e819-45d1-a91d-4bf1623a4324","Type":"ContainerDied","Data":"862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9"} Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.524837 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.543483 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.561348 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.565532 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.565623 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.565635 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.565660 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.565675 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:17Z","lastTransitionTime":"2025-11-24T17:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.584051 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.604988 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.628466 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.669482 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.669514 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.669526 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.669542 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.669554 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:17Z","lastTransitionTime":"2025-11-24T17:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.678281 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.693879 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.711578 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.731197 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.746406 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.756646 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.771415 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:17Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.772653 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.772692 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.772704 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.772721 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.772733 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:17Z","lastTransitionTime":"2025-11-24T17:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.875156 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.875196 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.875209 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.875225 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.875235 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:17Z","lastTransitionTime":"2025-11-24T17:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.978595 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.978648 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.978669 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.978696 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:17 crc kubenswrapper[4777]: I1124 17:01:17.978716 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:17Z","lastTransitionTime":"2025-11-24T17:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.081955 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.082048 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.082073 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.082102 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.082126 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:18Z","lastTransitionTime":"2025-11-24T17:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.188238 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.188369 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.188397 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.188428 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.188506 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:18Z","lastTransitionTime":"2025-11-24T17:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.244680 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.244778 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:18 crc kubenswrapper[4777]: E1124 17:01:18.244943 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:18 crc kubenswrapper[4777]: E1124 17:01:18.245128 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.291587 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.291617 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.291627 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.291642 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.291653 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:18Z","lastTransitionTime":"2025-11-24T17:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.394129 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.394180 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.394198 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.394230 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.394270 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:18Z","lastTransitionTime":"2025-11-24T17:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.497654 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.497718 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.497742 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.497770 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.497794 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:18Z","lastTransitionTime":"2025-11-24T17:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.516876 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" event={"ID":"cce07485-e819-45d1-a91d-4bf1623a4324","Type":"ContainerStarted","Data":"28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3"} Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.526511 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerStarted","Data":"614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb"} Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.528398 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.528470 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.540413 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.558835 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.566255 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.570757 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.580940 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.596672 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.600220 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.600261 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.600272 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.600290 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.600302 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:18Z","lastTransitionTime":"2025-11-24T17:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.610903 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.623186 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.637429 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.652625 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.666834 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.690218 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.702896 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.703006 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.703039 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.703050 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.703069 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.703081 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:18Z","lastTransitionTime":"2025-11-24T17:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.716961 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.732649 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.746338 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.758392 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.780767 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.797388 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.805449 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.805474 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.805482 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.805494 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.805503 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:18Z","lastTransitionTime":"2025-11-24T17:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.810909 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.823790 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.837802 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.847203 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.860091 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.874250 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.886474 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.891793 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.892002 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.892060 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:18 crc kubenswrapper[4777]: E1124 17:01:18.892088 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:01:26.892064625 +0000 UTC m=+45.051259674 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:01:18 crc kubenswrapper[4777]: E1124 17:01:18.892094 4777 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.892117 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:18 crc kubenswrapper[4777]: E1124 17:01:18.892159 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:26.892139527 +0000 UTC m=+45.051334656 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:01:18 crc kubenswrapper[4777]: E1124 17:01:18.892161 4777 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:01:18 crc kubenswrapper[4777]: E1124 17:01:18.892235 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:26.89221056 +0000 UTC m=+45.051405629 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:01:18 crc kubenswrapper[4777]: E1124 17:01:18.892253 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:01:18 crc kubenswrapper[4777]: E1124 17:01:18.892267 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:01:18 crc kubenswrapper[4777]: E1124 17:01:18.892277 4777 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:18 crc kubenswrapper[4777]: E1124 17:01:18.892312 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:26.892303132 +0000 UTC m=+45.051498181 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.903022 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.907012 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.907049 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.907060 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.907078 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.907090 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:18Z","lastTransitionTime":"2025-11-24T17:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.915674 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:18Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:18 crc kubenswrapper[4777]: I1124 17:01:18.993766 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:18 crc kubenswrapper[4777]: E1124 17:01:18.994017 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:01:18 crc kubenswrapper[4777]: E1124 17:01:18.994045 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:01:18 crc kubenswrapper[4777]: E1124 17:01:18.994065 4777 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:18 crc kubenswrapper[4777]: E1124 17:01:18.994145 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:26.994122871 +0000 UTC m=+45.153317960 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.010198 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.010244 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.010262 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.010285 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.010301 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:19Z","lastTransitionTime":"2025-11-24T17:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.113277 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.113926 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.113961 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.114011 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.114032 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:19Z","lastTransitionTime":"2025-11-24T17:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.218137 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.218209 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.218220 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.218242 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.218255 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:19Z","lastTransitionTime":"2025-11-24T17:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.244627 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:19 crc kubenswrapper[4777]: E1124 17:01:19.244832 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.321687 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.321760 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.321776 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.321800 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.321817 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:19Z","lastTransitionTime":"2025-11-24T17:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.424907 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.425018 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.425044 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.425078 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.425101 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:19Z","lastTransitionTime":"2025-11-24T17:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.528419 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.528477 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.528528 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.528554 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.528573 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:19Z","lastTransitionTime":"2025-11-24T17:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.529699 4777 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.631470 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.631514 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.631526 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.631545 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.631557 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:19Z","lastTransitionTime":"2025-11-24T17:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.734378 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.734434 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.734451 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.734477 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.734496 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:19Z","lastTransitionTime":"2025-11-24T17:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.838468 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.838494 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.838505 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.838520 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.838531 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:19Z","lastTransitionTime":"2025-11-24T17:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.941347 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.941391 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.941407 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.941433 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:19 crc kubenswrapper[4777]: I1124 17:01:19.941451 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:19Z","lastTransitionTime":"2025-11-24T17:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.044517 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.044573 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.044590 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.044612 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.044629 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:20Z","lastTransitionTime":"2025-11-24T17:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.147780 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.147816 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.147828 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.147843 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.147855 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:20Z","lastTransitionTime":"2025-11-24T17:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.244379 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:20 crc kubenswrapper[4777]: E1124 17:01:20.244522 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.244826 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:20 crc kubenswrapper[4777]: E1124 17:01:20.244948 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.251317 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.251347 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.251359 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.251373 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.251385 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:20Z","lastTransitionTime":"2025-11-24T17:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.354964 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.355070 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.355323 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.355358 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.355382 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:20Z","lastTransitionTime":"2025-11-24T17:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.460748 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.460800 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.460819 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.460844 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.460862 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:20Z","lastTransitionTime":"2025-11-24T17:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.534108 4777 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.563212 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.563241 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.563249 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.563262 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.563270 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:20Z","lastTransitionTime":"2025-11-24T17:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.666349 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.666392 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.666405 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.666425 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.666442 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:20Z","lastTransitionTime":"2025-11-24T17:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.770140 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.770231 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.770255 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.770286 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.770307 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:20Z","lastTransitionTime":"2025-11-24T17:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.909329 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.909363 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.909372 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.909390 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:20 crc kubenswrapper[4777]: I1124 17:01:20.909400 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:20Z","lastTransitionTime":"2025-11-24T17:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.012152 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.012217 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.012226 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.012245 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.012255 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:21Z","lastTransitionTime":"2025-11-24T17:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.114957 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.115129 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.115157 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.115196 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.115222 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:21Z","lastTransitionTime":"2025-11-24T17:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.218444 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.218509 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.218528 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.218553 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.218574 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:21Z","lastTransitionTime":"2025-11-24T17:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.244843 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:21 crc kubenswrapper[4777]: E1124 17:01:21.245076 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.255332 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn"] Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.255903 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.262840 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.263161 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.284172 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.302586 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.320891 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.320926 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.320938 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.320954 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.320989 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:21Z","lastTransitionTime":"2025-11-24T17:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.321291 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.347695 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/97bd214a-a09d-4bff-9cf3-5a6e1a4c863a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6v6jn\" (UID: \"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.347762 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/97bd214a-a09d-4bff-9cf3-5a6e1a4c863a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6v6jn\" (UID: \"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.347822 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/97bd214a-a09d-4bff-9cf3-5a6e1a4c863a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6v6jn\" (UID: \"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.347845 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wtnl\" (UniqueName: \"kubernetes.io/projected/97bd214a-a09d-4bff-9cf3-5a6e1a4c863a-kube-api-access-2wtnl\") pod \"ovnkube-control-plane-749d76644c-6v6jn\" (UID: \"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.348811 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.363993 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.383840 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.400817 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.416473 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.423527 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.423569 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.423582 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.423600 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.423612 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:21Z","lastTransitionTime":"2025-11-24T17:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.429663 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.448623 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.448840 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/97bd214a-a09d-4bff-9cf3-5a6e1a4c863a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6v6jn\" (UID: \"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.448900 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wtnl\" (UniqueName: \"kubernetes.io/projected/97bd214a-a09d-4bff-9cf3-5a6e1a4c863a-kube-api-access-2wtnl\") pod \"ovnkube-control-plane-749d76644c-6v6jn\" (UID: \"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.448936 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/97bd214a-a09d-4bff-9cf3-5a6e1a4c863a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6v6jn\" (UID: \"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.449741 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/97bd214a-a09d-4bff-9cf3-5a6e1a4c863a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6v6jn\" (UID: \"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.449817 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/97bd214a-a09d-4bff-9cf3-5a6e1a4c863a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6v6jn\" (UID: \"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.450212 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/97bd214a-a09d-4bff-9cf3-5a6e1a4c863a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6v6jn\" (UID: \"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.457409 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/97bd214a-a09d-4bff-9cf3-5a6e1a4c863a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6v6jn\" (UID: \"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.468495 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wtnl\" (UniqueName: \"kubernetes.io/projected/97bd214a-a09d-4bff-9cf3-5a6e1a4c863a-kube-api-access-2wtnl\") pod \"ovnkube-control-plane-749d76644c-6v6jn\" (UID: \"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.476917 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.493325 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.517170 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.526362 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.526384 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.526396 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.526412 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.526423 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:21Z","lastTransitionTime":"2025-11-24T17:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.540609 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.584713 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" Nov 24 17:01:21 crc kubenswrapper[4777]: W1124 17:01:21.601161 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97bd214a_a09d_4bff_9cf3_5a6e1a4c863a.slice/crio-83332d99355e437d09ef79592156b21f22da0af5054584a4a5f9272ec04f6ca2 WatchSource:0}: Error finding container 83332d99355e437d09ef79592156b21f22da0af5054584a4a5f9272ec04f6ca2: Status 404 returned error can't find the container with id 83332d99355e437d09ef79592156b21f22da0af5054584a4a5f9272ec04f6ca2 Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.632372 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.632411 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.632422 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.632438 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.632452 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:21Z","lastTransitionTime":"2025-11-24T17:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.735219 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.735294 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.735318 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.735351 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.735377 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:21Z","lastTransitionTime":"2025-11-24T17:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.839083 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.839132 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.839142 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.839160 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.839170 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:21Z","lastTransitionTime":"2025-11-24T17:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.943500 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.944126 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.944149 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.944182 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:21 crc kubenswrapper[4777]: I1124 17:01:21.944202 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:21Z","lastTransitionTime":"2025-11-24T17:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.047845 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.047906 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.047927 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.047951 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.047994 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:22Z","lastTransitionTime":"2025-11-24T17:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.150771 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.150827 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.150842 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.150865 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.150884 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:22Z","lastTransitionTime":"2025-11-24T17:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.244753 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.244776 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:22 crc kubenswrapper[4777]: E1124 17:01:22.245086 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:22 crc kubenswrapper[4777]: E1124 17:01:22.245169 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.253741 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.253773 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.253781 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.253795 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.253804 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:22Z","lastTransitionTime":"2025-11-24T17:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.356716 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.356776 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.356796 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.356823 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.356840 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:22Z","lastTransitionTime":"2025-11-24T17:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.459627 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.459674 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.459690 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.459712 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.459732 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:22Z","lastTransitionTime":"2025-11-24T17:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.545563 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/0.log" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.552842 4777 generic.go:334] "Generic (PLEG): container finished" podID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerID="614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb" exitCode=1 Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.552928 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerDied","Data":"614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb"} Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.555271 4777 scope.go:117] "RemoveContainer" containerID="614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.556709 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" event={"ID":"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a","Type":"ContainerStarted","Data":"b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160"} Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.556756 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" event={"ID":"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a","Type":"ContainerStarted","Data":"cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69"} Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.556777 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" event={"ID":"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a","Type":"ContainerStarted","Data":"83332d99355e437d09ef79592156b21f22da0af5054584a4a5f9272ec04f6ca2"} Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.562536 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.562771 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.562810 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.562839 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.562864 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:22Z","lastTransitionTime":"2025-11-24T17:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.578661 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.603064 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.628807 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.651819 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.666580 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.666634 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.666654 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.666680 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.666699 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:22Z","lastTransitionTime":"2025-11-24T17:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.672190 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.691051 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.717768 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.738293 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.758754 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.769863 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.769924 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.769948 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.770012 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.770039 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:22Z","lastTransitionTime":"2025-11-24T17:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.779339 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-kxqzn"] Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.780279 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:22 crc kubenswrapper[4777]: E1124 17:01:22.780415 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.782815 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.805708 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.822140 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.844955 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"message\\\":\\\"l\\\\nI1124 17:01:21.664518 6103 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 17:01:21.664583 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:01:21.664608 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:01:21.666581 6103 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 17:01:21.666609 6103 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 17:01:21.666662 6103 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 17:01:21.666671 6103 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 17:01:21.666744 6103 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 17:01:21.666727 6103 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 17:01:21.666769 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 17:01:21.666780 6103 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 17:01:21.666829 6103 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 17:01:21.666850 6103 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 17:01:21.667092 6103 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:01:21.667196 6103 factory.go:656] Stopping watch factory\\\\nI1124 17:01:21.667231 6103 ovnkube.go:599] Stopped ovnkube\\\\nI1124 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.858575 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.863991 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs\") pod \"network-metrics-daemon-kxqzn\" (UID: \"960e1bed-abef-4f45-9475-37eee6bd780a\") " pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.864034 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2vm5\" (UniqueName: \"kubernetes.io/projected/960e1bed-abef-4f45-9475-37eee6bd780a-kube-api-access-v2vm5\") pod \"network-metrics-daemon-kxqzn\" (UID: \"960e1bed-abef-4f45-9475-37eee6bd780a\") " pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.873321 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.873363 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.873380 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.873402 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.873418 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:22Z","lastTransitionTime":"2025-11-24T17:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.898894 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.921029 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.938980 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.953098 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.965176 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs\") pod \"network-metrics-daemon-kxqzn\" (UID: \"960e1bed-abef-4f45-9475-37eee6bd780a\") " pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.965272 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2vm5\" (UniqueName: \"kubernetes.io/projected/960e1bed-abef-4f45-9475-37eee6bd780a-kube-api-access-v2vm5\") pod \"network-metrics-daemon-kxqzn\" (UID: \"960e1bed-abef-4f45-9475-37eee6bd780a\") " pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:22 crc kubenswrapper[4777]: E1124 17:01:22.965348 4777 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:01:22 crc kubenswrapper[4777]: E1124 17:01:22.965424 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs podName:960e1bed-abef-4f45-9475-37eee6bd780a nodeName:}" failed. No retries permitted until 2025-11-24 17:01:23.46540499 +0000 UTC m=+41.624600159 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs") pod "network-metrics-daemon-kxqzn" (UID: "960e1bed-abef-4f45-9475-37eee6bd780a") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.967949 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.976494 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.976551 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.976565 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.976586 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.976625 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:22Z","lastTransitionTime":"2025-11-24T17:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.986776 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:22 crc kubenswrapper[4777]: I1124 17:01:22.988621 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2vm5\" (UniqueName: \"kubernetes.io/projected/960e1bed-abef-4f45-9475-37eee6bd780a-kube-api-access-v2vm5\") pod \"network-metrics-daemon-kxqzn\" (UID: \"960e1bed-abef-4f45-9475-37eee6bd780a\") " pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.022726 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"message\\\":\\\"l\\\\nI1124 17:01:21.664518 6103 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 17:01:21.664583 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:01:21.664608 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:01:21.666581 6103 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 17:01:21.666609 6103 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 17:01:21.666662 6103 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 17:01:21.666671 6103 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 17:01:21.666744 6103 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 17:01:21.666727 6103 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 17:01:21.666769 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 17:01:21.666780 6103 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 17:01:21.666829 6103 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 17:01:21.666850 6103 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 17:01:21.667092 6103 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:01:21.667196 6103 factory.go:656] Stopping watch factory\\\\nI1124 17:01:21.667231 6103 ovnkube.go:599] Stopped ovnkube\\\\nI1124 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.037056 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.049218 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.063093 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.078049 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.079397 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.079510 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.079531 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.079555 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.079576 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:23Z","lastTransitionTime":"2025-11-24T17:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.101652 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.121112 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.138126 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.157893 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.208167 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.208324 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.208348 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.208372 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.208390 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:23Z","lastTransitionTime":"2025-11-24T17:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.246801 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:23 crc kubenswrapper[4777]: E1124 17:01:23.247084 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.248911 4777 scope.go:117] "RemoveContainer" containerID="673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.277530 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.299855 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.311544 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.311606 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.311629 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.311658 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.311681 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:23Z","lastTransitionTime":"2025-11-24T17:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.327933 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.353863 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.373452 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.392745 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.411298 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.413929 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.414022 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.414048 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.414078 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.414102 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:23Z","lastTransitionTime":"2025-11-24T17:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.437760 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.468709 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.485266 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.503525 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.508003 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs\") pod \"network-metrics-daemon-kxqzn\" (UID: \"960e1bed-abef-4f45-9475-37eee6bd780a\") " pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:23 crc kubenswrapper[4777]: E1124 17:01:23.508217 4777 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:01:23 crc kubenswrapper[4777]: E1124 17:01:23.508303 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs podName:960e1bed-abef-4f45-9475-37eee6bd780a nodeName:}" failed. No retries permitted until 2025-11-24 17:01:24.508282502 +0000 UTC m=+42.667477571 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs") pod "network-metrics-daemon-kxqzn" (UID: "960e1bed-abef-4f45-9475-37eee6bd780a") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.520114 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.520174 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.520214 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.520247 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.520267 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:23Z","lastTransitionTime":"2025-11-24T17:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.535386 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.552595 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.576729 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/0.log" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.579010 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerStarted","Data":"3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b"} Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.579133 4777 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.586943 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"message\\\":\\\"l\\\\nI1124 17:01:21.664518 6103 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 17:01:21.664583 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:01:21.664608 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:01:21.666581 6103 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 17:01:21.666609 6103 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 17:01:21.666662 6103 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 17:01:21.666671 6103 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 17:01:21.666744 6103 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 17:01:21.666727 6103 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 17:01:21.666769 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 17:01:21.666780 6103 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 17:01:21.666829 6103 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 17:01:21.666850 6103 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 17:01:21.667092 6103 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:01:21.667196 6103 factory.go:656] Stopping watch factory\\\\nI1124 17:01:21.667231 6103 ovnkube.go:599] Stopped ovnkube\\\\nI1124 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.601193 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.619645 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"message\\\":\\\"l\\\\nI1124 17:01:21.664518 6103 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 17:01:21.664583 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:01:21.664608 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:01:21.666581 6103 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 17:01:21.666609 6103 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 17:01:21.666662 6103 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 17:01:21.666671 6103 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 17:01:21.666744 6103 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 17:01:21.666727 6103 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 17:01:21.666769 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 17:01:21.666780 6103 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 17:01:21.666829 6103 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 17:01:21.666850 6103 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 17:01:21.667092 6103 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:01:21.667196 6103 factory.go:656] Stopping watch factory\\\\nI1124 17:01:21.667231 6103 ovnkube.go:599] Stopped ovnkube\\\\nI1124 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.622013 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.622042 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.622051 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.622064 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.622073 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:23Z","lastTransitionTime":"2025-11-24T17:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.630362 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.643049 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.658327 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.671332 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.687312 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.702700 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.714629 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.724205 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.724251 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.724268 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.724289 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.724304 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:23Z","lastTransitionTime":"2025-11-24T17:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.731192 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.745823 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.761405 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.776253 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.790962 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.805589 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.823723 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.826450 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.826493 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.826526 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.826546 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.826558 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:23Z","lastTransitionTime":"2025-11-24T17:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.929563 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.929628 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.929645 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.929671 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:23 crc kubenswrapper[4777]: I1124 17:01:23.929703 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:23Z","lastTransitionTime":"2025-11-24T17:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.032158 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.032238 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.032255 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.032309 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.032326 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:24Z","lastTransitionTime":"2025-11-24T17:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.135593 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.135652 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.135670 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.135695 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.135714 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:24Z","lastTransitionTime":"2025-11-24T17:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.239656 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.239735 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.239761 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.239792 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.239812 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:24Z","lastTransitionTime":"2025-11-24T17:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.245091 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.245161 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.245243 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:24 crc kubenswrapper[4777]: E1124 17:01:24.245288 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:24 crc kubenswrapper[4777]: E1124 17:01:24.245450 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:24 crc kubenswrapper[4777]: E1124 17:01:24.245586 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.343923 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.343995 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.344012 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.344033 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.344051 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:24Z","lastTransitionTime":"2025-11-24T17:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.447326 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.447385 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.447395 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.447415 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.447428 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:24Z","lastTransitionTime":"2025-11-24T17:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.519365 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs\") pod \"network-metrics-daemon-kxqzn\" (UID: \"960e1bed-abef-4f45-9475-37eee6bd780a\") " pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:24 crc kubenswrapper[4777]: E1124 17:01:24.519582 4777 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:01:24 crc kubenswrapper[4777]: E1124 17:01:24.519710 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs podName:960e1bed-abef-4f45-9475-37eee6bd780a nodeName:}" failed. No retries permitted until 2025-11-24 17:01:26.519679751 +0000 UTC m=+44.678874830 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs") pod "network-metrics-daemon-kxqzn" (UID: "960e1bed-abef-4f45-9475-37eee6bd780a") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.550545 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.550600 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.550618 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.550645 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.550665 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:24Z","lastTransitionTime":"2025-11-24T17:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.592890 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.596663 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.597261 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.600891 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/1.log" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.603545 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/0.log" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.608363 4777 generic.go:334] "Generic (PLEG): container finished" podID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerID="3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b" exitCode=1 Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.608435 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerDied","Data":"3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.608497 4777 scope.go:117] "RemoveContainer" containerID="614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.610148 4777 scope.go:117] "RemoveContainer" containerID="3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b" Nov 24 17:01:24 crc kubenswrapper[4777]: E1124 17:01:24.610432 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\"" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.627309 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.644536 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.653438 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.653481 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.653492 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.653509 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.653519 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:24Z","lastTransitionTime":"2025-11-24T17:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.662256 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.682688 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.712666 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"message\\\":\\\"l\\\\nI1124 17:01:21.664518 6103 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 17:01:21.664583 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:01:21.664608 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:01:21.666581 6103 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 17:01:21.666609 6103 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 17:01:21.666662 6103 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 17:01:21.666671 6103 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 17:01:21.666744 6103 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 17:01:21.666727 6103 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 17:01:21.666769 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 17:01:21.666780 6103 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 17:01:21.666829 6103 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 17:01:21.666850 6103 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 17:01:21.667092 6103 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:01:21.667196 6103 factory.go:656] Stopping watch factory\\\\nI1124 17:01:21.667231 6103 ovnkube.go:599] Stopped ovnkube\\\\nI1124 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.727477 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.732243 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.732329 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.732347 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.732402 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.732420 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:24Z","lastTransitionTime":"2025-11-24T17:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.742656 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: E1124 17:01:24.750008 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.755310 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.755502 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.755670 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.755816 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.755946 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:24Z","lastTransitionTime":"2025-11-24T17:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.757191 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.772227 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: E1124 17:01:24.775326 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.779226 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.779303 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.779321 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.779373 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.779391 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:24Z","lastTransitionTime":"2025-11-24T17:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.786229 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: E1124 17:01:24.799516 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.801216 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.803721 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.803752 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.803761 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.803775 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.803786 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:24Z","lastTransitionTime":"2025-11-24T17:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.814486 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: E1124 17:01:24.819924 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.824257 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.824286 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.824298 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.824315 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.824328 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:24Z","lastTransitionTime":"2025-11-24T17:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.833306 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: E1124 17:01:24.843266 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: E1124 17:01:24.843449 4777 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.848114 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.848143 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.848155 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.848171 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.848186 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:24Z","lastTransitionTime":"2025-11-24T17:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.855795 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.872518 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.886201 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.897214 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.917723 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"message\\\":\\\"l\\\\nI1124 17:01:21.664518 6103 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 17:01:21.664583 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:01:21.664608 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:01:21.666581 6103 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 17:01:21.666609 6103 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 17:01:21.666662 6103 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 17:01:21.666671 6103 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 17:01:21.666744 6103 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 17:01:21.666727 6103 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 17:01:21.666769 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 17:01:21.666780 6103 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 17:01:21.666829 6103 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 17:01:21.666850 6103 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 17:01:21.667092 6103 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:01:21.667196 6103 factory.go:656] Stopping watch factory\\\\nI1124 17:01:21.667231 6103 ovnkube.go:599] Stopped ovnkube\\\\nI1124 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"gressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 17:01:23.957314 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:01:23.957339 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:01:23.957286 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:01:23.957410 6297 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 17:01:23.957443 6297 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 17:01:23.957468 6297 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 17:01:23.957495 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 17:01:23.957530 6297 factory.go:656] Stopping watch factory\\\\nI1124 17:01:23.957556 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 17:01:23.957663 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 17:01:23.957688 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 17:01:23.957711 6297 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 17:01:23.957734 6297 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 17:01:23.957818 6297 ovnkube.go:599] Stopped ovnkube\\\\nI1124 17:01:23.957874 6297 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 17:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.931091 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.946026 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.950945 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.950998 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.951011 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.951029 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.951042 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:24Z","lastTransitionTime":"2025-11-24T17:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.964041 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.980511 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:24 crc kubenswrapper[4777]: I1124 17:01:24.996812 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:24Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.015386 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:25Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.030300 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:25Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.047833 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:25Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.053303 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.053335 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.053348 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.053364 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.053376 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:25Z","lastTransitionTime":"2025-11-24T17:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.067898 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:25Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.080684 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:25Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.094515 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:25Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.105623 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:25Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.156335 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.156379 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.156388 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.156404 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.156413 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:25Z","lastTransitionTime":"2025-11-24T17:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.244887 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:25 crc kubenswrapper[4777]: E1124 17:01:25.245065 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.259298 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.259362 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.259385 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.259410 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.259428 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:25Z","lastTransitionTime":"2025-11-24T17:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.362925 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.363012 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.363036 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.363065 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.363086 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:25Z","lastTransitionTime":"2025-11-24T17:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.465291 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.465357 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.465374 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.465401 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.465421 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:25Z","lastTransitionTime":"2025-11-24T17:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.568654 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.568736 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.568760 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.568791 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.568814 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:25Z","lastTransitionTime":"2025-11-24T17:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.613948 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/1.log" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.671140 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.671162 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.671172 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.671185 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.671193 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:25Z","lastTransitionTime":"2025-11-24T17:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.773705 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.773753 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.773768 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.773794 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.773817 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:25Z","lastTransitionTime":"2025-11-24T17:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.876713 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.876786 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.876812 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.876849 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.876873 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:25Z","lastTransitionTime":"2025-11-24T17:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.980404 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.980474 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.980493 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.980523 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:25 crc kubenswrapper[4777]: I1124 17:01:25.980542 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:25Z","lastTransitionTime":"2025-11-24T17:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.083201 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.083280 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.083304 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.083341 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.083367 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:26Z","lastTransitionTime":"2025-11-24T17:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.186852 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.187426 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.187443 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.187474 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.187496 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:26Z","lastTransitionTime":"2025-11-24T17:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.245200 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.245556 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.245817 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:26 crc kubenswrapper[4777]: E1124 17:01:26.245808 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:26 crc kubenswrapper[4777]: E1124 17:01:26.246005 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:26 crc kubenswrapper[4777]: E1124 17:01:26.246320 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.292743 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.292797 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.292816 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.292840 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.292859 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:26Z","lastTransitionTime":"2025-11-24T17:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.395534 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.395590 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.395613 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.395639 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.395661 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:26Z","lastTransitionTime":"2025-11-24T17:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.498444 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.498501 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.498519 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.498549 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.498567 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:26Z","lastTransitionTime":"2025-11-24T17:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.545330 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs\") pod \"network-metrics-daemon-kxqzn\" (UID: \"960e1bed-abef-4f45-9475-37eee6bd780a\") " pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:26 crc kubenswrapper[4777]: E1124 17:01:26.545643 4777 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:01:26 crc kubenswrapper[4777]: E1124 17:01:26.545747 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs podName:960e1bed-abef-4f45-9475-37eee6bd780a nodeName:}" failed. No retries permitted until 2025-11-24 17:01:30.545718168 +0000 UTC m=+48.704913257 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs") pod "network-metrics-daemon-kxqzn" (UID: "960e1bed-abef-4f45-9475-37eee6bd780a") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.601105 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.601196 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.601217 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.601242 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.601289 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:26Z","lastTransitionTime":"2025-11-24T17:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.704719 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.704759 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.704769 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.704782 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.704792 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:26Z","lastTransitionTime":"2025-11-24T17:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.808040 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.808127 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.808152 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.808185 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.808209 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:26Z","lastTransitionTime":"2025-11-24T17:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.911095 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.911132 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.911141 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.911156 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.911167 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:26Z","lastTransitionTime":"2025-11-24T17:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.950131 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.950265 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:26 crc kubenswrapper[4777]: E1124 17:01:26.950309 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:01:42.950271715 +0000 UTC m=+61.109466804 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:01:26 crc kubenswrapper[4777]: E1124 17:01:26.950439 4777 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.950442 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:26 crc kubenswrapper[4777]: E1124 17:01:26.950499 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:42.950482232 +0000 UTC m=+61.109677331 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:01:26 crc kubenswrapper[4777]: E1124 17:01:26.950561 4777 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:01:26 crc kubenswrapper[4777]: I1124 17:01:26.950570 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:26 crc kubenswrapper[4777]: E1124 17:01:26.950615 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:42.950600656 +0000 UTC m=+61.109795735 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:01:26 crc kubenswrapper[4777]: E1124 17:01:26.950716 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:01:26 crc kubenswrapper[4777]: E1124 17:01:26.950736 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:01:26 crc kubenswrapper[4777]: E1124 17:01:26.950750 4777 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:26 crc kubenswrapper[4777]: E1124 17:01:26.950794 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:42.950782871 +0000 UTC m=+61.109978010 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.013860 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.013912 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.013929 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.013953 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.014012 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:27Z","lastTransitionTime":"2025-11-24T17:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.052302 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:27 crc kubenswrapper[4777]: E1124 17:01:27.052542 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:01:27 crc kubenswrapper[4777]: E1124 17:01:27.052567 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:01:27 crc kubenswrapper[4777]: E1124 17:01:27.052581 4777 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:27 crc kubenswrapper[4777]: E1124 17:01:27.052641 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:01:43.05262396 +0000 UTC m=+61.211819019 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.116334 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.116392 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.116412 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.116436 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.116453 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:27Z","lastTransitionTime":"2025-11-24T17:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.218873 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.218922 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.218939 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.218961 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.219003 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:27Z","lastTransitionTime":"2025-11-24T17:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.245057 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:27 crc kubenswrapper[4777]: E1124 17:01:27.245274 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.321507 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.321546 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.321556 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.321572 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.321584 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:27Z","lastTransitionTime":"2025-11-24T17:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.425231 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.425310 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.425330 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.425357 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.425770 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:27Z","lastTransitionTime":"2025-11-24T17:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.529642 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.529782 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.529808 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.529836 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.529860 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:27Z","lastTransitionTime":"2025-11-24T17:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.632166 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.632231 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.632257 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.632286 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.632309 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:27Z","lastTransitionTime":"2025-11-24T17:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.735678 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.735754 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.735781 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.735816 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.735838 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:27Z","lastTransitionTime":"2025-11-24T17:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.839070 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.839138 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.839160 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.839189 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.839212 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:27Z","lastTransitionTime":"2025-11-24T17:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.942108 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.942152 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.942169 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.942190 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:27 crc kubenswrapper[4777]: I1124 17:01:27.942207 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:27Z","lastTransitionTime":"2025-11-24T17:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.045454 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.045528 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.045545 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.045572 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.045589 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:28Z","lastTransitionTime":"2025-11-24T17:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.148822 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.148887 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.148905 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.148930 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.148949 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:28Z","lastTransitionTime":"2025-11-24T17:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.244240 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.244321 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.244275 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:28 crc kubenswrapper[4777]: E1124 17:01:28.244488 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:28 crc kubenswrapper[4777]: E1124 17:01:28.244842 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:28 crc kubenswrapper[4777]: E1124 17:01:28.244720 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.253590 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.253630 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.253642 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.253657 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.253671 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:28Z","lastTransitionTime":"2025-11-24T17:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.357087 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.357185 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.357211 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.357246 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.357269 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:28Z","lastTransitionTime":"2025-11-24T17:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.460623 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.460786 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.460851 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.460883 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.460941 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:28Z","lastTransitionTime":"2025-11-24T17:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.564423 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.565266 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.565292 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.565330 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.565359 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:28Z","lastTransitionTime":"2025-11-24T17:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.671010 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.671058 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.671071 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.671096 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.671112 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:28Z","lastTransitionTime":"2025-11-24T17:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.773992 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.774061 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.774072 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.774092 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.774107 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:28Z","lastTransitionTime":"2025-11-24T17:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.877138 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.877180 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.877188 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.877202 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.877212 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:28Z","lastTransitionTime":"2025-11-24T17:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.980606 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.980665 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.980689 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.980729 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:28 crc kubenswrapper[4777]: I1124 17:01:28.980748 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:28Z","lastTransitionTime":"2025-11-24T17:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.088048 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.088081 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.088089 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.088104 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.088113 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:29Z","lastTransitionTime":"2025-11-24T17:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.190408 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.190452 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.190464 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.190485 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.190497 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:29Z","lastTransitionTime":"2025-11-24T17:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.244779 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:29 crc kubenswrapper[4777]: E1124 17:01:29.245013 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.293045 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.293104 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.293117 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.293135 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.293146 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:29Z","lastTransitionTime":"2025-11-24T17:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.395213 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.395291 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.395315 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.395347 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.395370 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:29Z","lastTransitionTime":"2025-11-24T17:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.498457 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.498513 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.498531 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.498558 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.498579 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:29Z","lastTransitionTime":"2025-11-24T17:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.601474 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.601516 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.601525 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.601539 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.601551 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:29Z","lastTransitionTime":"2025-11-24T17:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.703941 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.704016 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.704029 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.704044 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.704054 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:29Z","lastTransitionTime":"2025-11-24T17:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.807019 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.807127 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.807142 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.807155 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.807166 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:29Z","lastTransitionTime":"2025-11-24T17:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.909030 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.909089 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.909107 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.909132 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:29 crc kubenswrapper[4777]: I1124 17:01:29.909150 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:29Z","lastTransitionTime":"2025-11-24T17:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.012611 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.012659 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.012677 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.012699 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.012716 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:30Z","lastTransitionTime":"2025-11-24T17:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.116421 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.116485 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.116501 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.116523 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.116537 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:30Z","lastTransitionTime":"2025-11-24T17:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.221880 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.221922 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.221936 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.221958 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.221975 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:30Z","lastTransitionTime":"2025-11-24T17:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.244566 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.244643 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.244594 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:30 crc kubenswrapper[4777]: E1124 17:01:30.244774 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:30 crc kubenswrapper[4777]: E1124 17:01:30.245019 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:30 crc kubenswrapper[4777]: E1124 17:01:30.245121 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.324182 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.324222 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.324234 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.324445 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.324458 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:30Z","lastTransitionTime":"2025-11-24T17:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.427581 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.427646 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.427664 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.427692 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.427712 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:30Z","lastTransitionTime":"2025-11-24T17:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.530890 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.530949 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.530982 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.531051 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.531077 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:30Z","lastTransitionTime":"2025-11-24T17:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.625060 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs\") pod \"network-metrics-daemon-kxqzn\" (UID: \"960e1bed-abef-4f45-9475-37eee6bd780a\") " pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:30 crc kubenswrapper[4777]: E1124 17:01:30.625311 4777 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:01:30 crc kubenswrapper[4777]: E1124 17:01:30.625442 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs podName:960e1bed-abef-4f45-9475-37eee6bd780a nodeName:}" failed. No retries permitted until 2025-11-24 17:01:38.625413363 +0000 UTC m=+56.784608452 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs") pod "network-metrics-daemon-kxqzn" (UID: "960e1bed-abef-4f45-9475-37eee6bd780a") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.634383 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.634428 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.634447 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.634471 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.634488 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:30Z","lastTransitionTime":"2025-11-24T17:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.736846 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.736905 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.736923 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.736946 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.737005 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:30Z","lastTransitionTime":"2025-11-24T17:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.841409 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.841505 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.841523 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.841550 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.841570 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:30Z","lastTransitionTime":"2025-11-24T17:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.944618 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.944671 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.944694 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.944718 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:30 crc kubenswrapper[4777]: I1124 17:01:30.944737 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:30Z","lastTransitionTime":"2025-11-24T17:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.048401 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.048464 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.048481 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.048507 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.048525 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:31Z","lastTransitionTime":"2025-11-24T17:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.151718 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.151786 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.151808 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.151836 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.151859 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:31Z","lastTransitionTime":"2025-11-24T17:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.244313 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:31 crc kubenswrapper[4777]: E1124 17:01:31.244572 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.254518 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.254801 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.255031 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.255250 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.255443 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:31Z","lastTransitionTime":"2025-11-24T17:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.359004 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.359054 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.359070 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.359094 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.359116 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:31Z","lastTransitionTime":"2025-11-24T17:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.461695 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.461749 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.461765 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.461789 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.461807 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:31Z","lastTransitionTime":"2025-11-24T17:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.570108 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.570160 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.570177 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.570234 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.570250 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:31Z","lastTransitionTime":"2025-11-24T17:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.673620 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.673663 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.673673 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.673688 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.673699 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:31Z","lastTransitionTime":"2025-11-24T17:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.777109 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.777174 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.777191 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.777218 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.777236 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:31Z","lastTransitionTime":"2025-11-24T17:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.880041 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.880102 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.880119 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.880142 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.880160 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:31Z","lastTransitionTime":"2025-11-24T17:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.982952 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.983015 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.983028 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.983044 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:31 crc kubenswrapper[4777]: I1124 17:01:31.983055 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:31Z","lastTransitionTime":"2025-11-24T17:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.086059 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.086124 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.086157 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.086195 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.086218 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:32Z","lastTransitionTime":"2025-11-24T17:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.189616 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.189652 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.189663 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.189678 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.189689 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:32Z","lastTransitionTime":"2025-11-24T17:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.244385 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.244385 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:32 crc kubenswrapper[4777]: E1124 17:01:32.244576 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:32 crc kubenswrapper[4777]: E1124 17:01:32.244688 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.245159 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:32 crc kubenswrapper[4777]: E1124 17:01:32.245349 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.292900 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.292963 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.293016 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.293044 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.293066 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:32Z","lastTransitionTime":"2025-11-24T17:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.395523 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.395577 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.395595 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.395618 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.395637 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:32Z","lastTransitionTime":"2025-11-24T17:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.498762 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.498798 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.498806 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.498819 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.498828 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:32Z","lastTransitionTime":"2025-11-24T17:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.601311 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.601375 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.601398 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.601426 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.601447 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:32Z","lastTransitionTime":"2025-11-24T17:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.704650 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.704680 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.704688 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.704700 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.704710 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:32Z","lastTransitionTime":"2025-11-24T17:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.808119 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.808217 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.808249 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.808283 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.808306 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:32Z","lastTransitionTime":"2025-11-24T17:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.912163 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.912243 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.912263 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.912293 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:32 crc kubenswrapper[4777]: I1124 17:01:32.912314 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:32Z","lastTransitionTime":"2025-11-24T17:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.014838 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.014917 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.014940 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.015017 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.015038 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:33Z","lastTransitionTime":"2025-11-24T17:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.117815 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.117879 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.117896 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.117922 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.117940 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:33Z","lastTransitionTime":"2025-11-24T17:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.221076 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.221148 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.221168 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.221201 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.221223 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:33Z","lastTransitionTime":"2025-11-24T17:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.244485 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:33 crc kubenswrapper[4777]: E1124 17:01:33.244685 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.267657 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.284401 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.315860 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://614d904d9c15711917c204508b3f6158b24f311d945a386876fcb0596ca62beb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"message\\\":\\\"l\\\\nI1124 17:01:21.664518 6103 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 17:01:21.664583 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:01:21.664608 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:01:21.666581 6103 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 17:01:21.666609 6103 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 17:01:21.666662 6103 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 17:01:21.666671 6103 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 17:01:21.666744 6103 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 17:01:21.666727 6103 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 17:01:21.666769 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 17:01:21.666780 6103 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 17:01:21.666829 6103 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 17:01:21.666850 6103 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 17:01:21.667092 6103 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:01:21.667196 6103 factory.go:656] Stopping watch factory\\\\nI1124 17:01:21.667231 6103 ovnkube.go:599] Stopped ovnkube\\\\nI1124 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"gressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 17:01:23.957314 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:01:23.957339 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:01:23.957286 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:01:23.957410 6297 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 17:01:23.957443 6297 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 17:01:23.957468 6297 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 17:01:23.957495 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 17:01:23.957530 6297 factory.go:656] Stopping watch factory\\\\nI1124 17:01:23.957556 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 17:01:23.957663 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 17:01:23.957688 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 17:01:23.957711 6297 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 17:01:23.957734 6297 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 17:01:23.957818 6297 ovnkube.go:599] Stopped ovnkube\\\\nI1124 17:01:23.957874 6297 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 17:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.324945 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.325009 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.325022 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.325038 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.325048 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:33Z","lastTransitionTime":"2025-11-24T17:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.332083 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.349742 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.367364 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.388325 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.415490 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.427326 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.427392 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.427415 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.427447 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.427475 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:33Z","lastTransitionTime":"2025-11-24T17:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.439910 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.453666 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.455277 4777 scope.go:117] "RemoveContainer" containerID="3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b" Nov 24 17:01:33 crc kubenswrapper[4777]: E1124 17:01:33.455551 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\"" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.459386 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.479151 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.503760 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.528558 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.531250 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.531319 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.531339 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.531366 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.531387 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:33Z","lastTransitionTime":"2025-11-24T17:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.552047 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.577459 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.600705 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.619510 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.634892 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.634925 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.634935 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.634953 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.634976 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:33Z","lastTransitionTime":"2025-11-24T17:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.641534 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.663093 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.686817 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.705267 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.738028 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.738203 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.738218 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.738236 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.738248 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:33Z","lastTransitionTime":"2025-11-24T17:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.740306 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"gressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 17:01:23.957314 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:01:23.957339 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:01:23.957286 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:01:23.957410 6297 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 17:01:23.957443 6297 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 17:01:23.957468 6297 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 17:01:23.957495 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 17:01:23.957530 6297 factory.go:656] Stopping watch factory\\\\nI1124 17:01:23.957556 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 17:01:23.957663 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 17:01:23.957688 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 17:01:23.957711 6297 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 17:01:23.957734 6297 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 17:01:23.957818 6297 ovnkube.go:599] Stopped ovnkube\\\\nI1124 17:01:23.957874 6297 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 17:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.760602 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.783428 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.804697 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.830488 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.841255 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.841317 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.841336 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.841361 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.841379 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:33Z","lastTransitionTime":"2025-11-24T17:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.858518 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.872446 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.892820 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.914299 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.944593 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.944654 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.944674 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.944700 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:33 crc kubenswrapper[4777]: I1124 17:01:33.944722 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:33Z","lastTransitionTime":"2025-11-24T17:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.049600 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.049667 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.049684 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.049709 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.049729 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:34Z","lastTransitionTime":"2025-11-24T17:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.152478 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.152542 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.152561 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.152586 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.152607 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:34Z","lastTransitionTime":"2025-11-24T17:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.244610 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.244610 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:34 crc kubenswrapper[4777]: E1124 17:01:34.244781 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.245009 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:34 crc kubenswrapper[4777]: E1124 17:01:34.245010 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:34 crc kubenswrapper[4777]: E1124 17:01:34.245476 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.255782 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.255841 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.256054 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.256094 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.256115 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:34Z","lastTransitionTime":"2025-11-24T17:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.358824 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.358901 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.358922 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.358950 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.359000 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:34Z","lastTransitionTime":"2025-11-24T17:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.462919 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.463017 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.463047 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.463078 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.463100 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:34Z","lastTransitionTime":"2025-11-24T17:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.565554 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.565599 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.565616 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.565637 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.565652 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:34Z","lastTransitionTime":"2025-11-24T17:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.668767 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.668865 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.668883 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.668908 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.668929 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:34Z","lastTransitionTime":"2025-11-24T17:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.771627 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.771729 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.771754 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.771785 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.771806 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:34Z","lastTransitionTime":"2025-11-24T17:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.874772 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.874835 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.874851 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.874876 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.874894 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:34Z","lastTransitionTime":"2025-11-24T17:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.977734 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.977795 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.977822 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.977850 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:34 crc kubenswrapper[4777]: I1124 17:01:34.977873 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:34Z","lastTransitionTime":"2025-11-24T17:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.081627 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.081703 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.081725 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.081757 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.081778 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:35Z","lastTransitionTime":"2025-11-24T17:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.090207 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.090258 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.090280 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.090306 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.090326 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:35Z","lastTransitionTime":"2025-11-24T17:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:35 crc kubenswrapper[4777]: E1124 17:01:35.111728 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.116531 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.116607 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.116630 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.116662 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.116685 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:35Z","lastTransitionTime":"2025-11-24T17:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:35 crc kubenswrapper[4777]: E1124 17:01:35.136922 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.142167 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.142223 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.142241 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.142263 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.142280 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:35Z","lastTransitionTime":"2025-11-24T17:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:35 crc kubenswrapper[4777]: E1124 17:01:35.162675 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.169149 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.169234 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.169259 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.169360 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.169389 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:35Z","lastTransitionTime":"2025-11-24T17:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:35 crc kubenswrapper[4777]: E1124 17:01:35.226091 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.235392 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.235449 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.235470 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.235495 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.235516 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:35Z","lastTransitionTime":"2025-11-24T17:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.244946 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:35 crc kubenswrapper[4777]: E1124 17:01:35.245168 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:35 crc kubenswrapper[4777]: E1124 17:01:35.256590 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: E1124 17:01:35.256792 4777 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.258911 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.258962 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.258989 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.259005 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.259015 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:35Z","lastTransitionTime":"2025-11-24T17:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.361955 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.362063 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.362085 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.362118 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.362144 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:35Z","lastTransitionTime":"2025-11-24T17:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.457412 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.466015 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.466078 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.466101 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.466129 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.466148 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:35Z","lastTransitionTime":"2025-11-24T17:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.476900 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.487790 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.510655 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.530918 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.554081 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.569776 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.569870 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.569894 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.569918 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.569936 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:35Z","lastTransitionTime":"2025-11-24T17:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.575928 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.594272 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.633462 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"gressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 17:01:23.957314 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:01:23.957339 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:01:23.957286 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:01:23.957410 6297 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 17:01:23.957443 6297 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 17:01:23.957468 6297 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 17:01:23.957495 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 17:01:23.957530 6297 factory.go:656] Stopping watch factory\\\\nI1124 17:01:23.957556 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 17:01:23.957663 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 17:01:23.957688 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 17:01:23.957711 6297 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 17:01:23.957734 6297 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 17:01:23.957818 6297 ovnkube.go:599] Stopped ovnkube\\\\nI1124 17:01:23.957874 6297 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 17:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.653618 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.672171 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.675035 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.675142 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.675170 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.675232 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.675251 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:35Z","lastTransitionTime":"2025-11-24T17:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.698248 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.723107 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.751523 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.772956 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.778429 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.778502 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.778529 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.778566 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.778594 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:35Z","lastTransitionTime":"2025-11-24T17:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.791842 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.811620 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:35Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.881093 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.881150 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.881168 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.881194 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.881210 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:35Z","lastTransitionTime":"2025-11-24T17:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.984421 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.984501 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.984525 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.984555 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:35 crc kubenswrapper[4777]: I1124 17:01:35.984579 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:35Z","lastTransitionTime":"2025-11-24T17:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.087361 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.087418 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.087445 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.087473 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.087490 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:36Z","lastTransitionTime":"2025-11-24T17:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.190731 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.190810 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.190832 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.190857 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.190878 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:36Z","lastTransitionTime":"2025-11-24T17:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.244544 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.244662 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:36 crc kubenswrapper[4777]: E1124 17:01:36.244785 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.244545 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:36 crc kubenswrapper[4777]: E1124 17:01:36.244938 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:36 crc kubenswrapper[4777]: E1124 17:01:36.245077 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.294261 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.294336 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.294356 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.294378 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.294396 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:36Z","lastTransitionTime":"2025-11-24T17:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.398237 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.398295 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.398312 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.398336 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.398353 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:36Z","lastTransitionTime":"2025-11-24T17:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.501657 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.501716 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.501736 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.501760 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.501777 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:36Z","lastTransitionTime":"2025-11-24T17:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.604939 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.605046 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.605064 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.605088 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.605107 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:36Z","lastTransitionTime":"2025-11-24T17:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.708238 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.708564 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.708737 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.708889 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.709111 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:36Z","lastTransitionTime":"2025-11-24T17:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.811596 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.811694 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.811715 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.811739 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.811756 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:36Z","lastTransitionTime":"2025-11-24T17:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.914412 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.914717 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.914915 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.915203 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:36 crc kubenswrapper[4777]: I1124 17:01:36.915400 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:36Z","lastTransitionTime":"2025-11-24T17:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.018455 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.019327 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.019515 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.019644 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.019769 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:37Z","lastTransitionTime":"2025-11-24T17:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.122530 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.122942 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.123205 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.123362 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.123488 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:37Z","lastTransitionTime":"2025-11-24T17:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.225735 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.225800 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.225812 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.225851 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.225863 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:37Z","lastTransitionTime":"2025-11-24T17:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.244345 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:37 crc kubenswrapper[4777]: E1124 17:01:37.244453 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.328074 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.328133 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.328150 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.328177 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.328197 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:37Z","lastTransitionTime":"2025-11-24T17:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.430722 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.431108 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.431301 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.431512 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.431690 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:37Z","lastTransitionTime":"2025-11-24T17:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.534681 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.534748 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.534765 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.534791 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.534810 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:37Z","lastTransitionTime":"2025-11-24T17:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.640228 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.640331 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.640352 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.640383 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.640405 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:37Z","lastTransitionTime":"2025-11-24T17:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.743189 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.743245 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.743259 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.743280 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.743298 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:37Z","lastTransitionTime":"2025-11-24T17:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.845438 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.845677 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.845738 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.845818 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.845873 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:37Z","lastTransitionTime":"2025-11-24T17:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.949323 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.949357 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.949366 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.949381 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:37 crc kubenswrapper[4777]: I1124 17:01:37.949392 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:37Z","lastTransitionTime":"2025-11-24T17:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.051335 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.051369 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.051380 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.051394 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.051404 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:38Z","lastTransitionTime":"2025-11-24T17:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.154182 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.154263 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.154289 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.154322 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.154342 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:38Z","lastTransitionTime":"2025-11-24T17:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.245142 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.245202 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.245142 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:38 crc kubenswrapper[4777]: E1124 17:01:38.245352 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:38 crc kubenswrapper[4777]: E1124 17:01:38.245428 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:38 crc kubenswrapper[4777]: E1124 17:01:38.245567 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.256705 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.256760 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.256777 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.256801 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.256820 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:38Z","lastTransitionTime":"2025-11-24T17:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.360657 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.360710 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.360721 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.360738 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.360750 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:38Z","lastTransitionTime":"2025-11-24T17:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.464038 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.464098 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.464144 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.464178 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.464202 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:38Z","lastTransitionTime":"2025-11-24T17:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.566670 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.566824 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.566863 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.566893 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.566915 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:38Z","lastTransitionTime":"2025-11-24T17:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.640181 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs\") pod \"network-metrics-daemon-kxqzn\" (UID: \"960e1bed-abef-4f45-9475-37eee6bd780a\") " pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:38 crc kubenswrapper[4777]: E1124 17:01:38.640399 4777 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:01:38 crc kubenswrapper[4777]: E1124 17:01:38.640467 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs podName:960e1bed-abef-4f45-9475-37eee6bd780a nodeName:}" failed. No retries permitted until 2025-11-24 17:01:54.640447106 +0000 UTC m=+72.799642155 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs") pod "network-metrics-daemon-kxqzn" (UID: "960e1bed-abef-4f45-9475-37eee6bd780a") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.669017 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.669087 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.669104 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.669129 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.669145 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:38Z","lastTransitionTime":"2025-11-24T17:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.771520 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.771863 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.771881 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.771918 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.771941 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:38Z","lastTransitionTime":"2025-11-24T17:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.874722 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.874786 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.874803 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.874826 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.874844 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:38Z","lastTransitionTime":"2025-11-24T17:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.978672 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.978768 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.978783 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.978801 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:38 crc kubenswrapper[4777]: I1124 17:01:38.978813 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:38Z","lastTransitionTime":"2025-11-24T17:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.082579 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.082649 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.082668 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.082699 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.082719 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:39Z","lastTransitionTime":"2025-11-24T17:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.187340 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.187467 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.187488 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.187519 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.187545 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:39Z","lastTransitionTime":"2025-11-24T17:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.245335 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:39 crc kubenswrapper[4777]: E1124 17:01:39.245536 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.291059 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.291120 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.291134 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.291154 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.291168 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:39Z","lastTransitionTime":"2025-11-24T17:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.394615 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.394680 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.394689 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.394708 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.394721 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:39Z","lastTransitionTime":"2025-11-24T17:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.498487 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.498568 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.498602 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.498634 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.498653 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:39Z","lastTransitionTime":"2025-11-24T17:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.602426 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.602515 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.602541 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.602569 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.602589 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:39Z","lastTransitionTime":"2025-11-24T17:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.705233 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.705294 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.705312 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.705335 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.705350 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:39Z","lastTransitionTime":"2025-11-24T17:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.809063 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.809125 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.809143 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.809166 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.809183 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:39Z","lastTransitionTime":"2025-11-24T17:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.912062 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.912112 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.912128 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.912152 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:39 crc kubenswrapper[4777]: I1124 17:01:39.912168 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:39Z","lastTransitionTime":"2025-11-24T17:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.015543 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.015639 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.015657 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.015683 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.015702 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:40Z","lastTransitionTime":"2025-11-24T17:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.119081 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.119148 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.119166 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.119190 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.119207 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:40Z","lastTransitionTime":"2025-11-24T17:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.223055 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.223142 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.223166 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.223198 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.223221 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:40Z","lastTransitionTime":"2025-11-24T17:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.244299 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.244403 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.244320 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:40 crc kubenswrapper[4777]: E1124 17:01:40.244550 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:40 crc kubenswrapper[4777]: E1124 17:01:40.244685 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:40 crc kubenswrapper[4777]: E1124 17:01:40.244819 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.332854 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.332938 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.333019 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.333054 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.333079 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:40Z","lastTransitionTime":"2025-11-24T17:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.439584 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.439779 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.439799 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.439826 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.439887 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:40Z","lastTransitionTime":"2025-11-24T17:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.542759 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.542836 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.542855 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.542879 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.542933 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:40Z","lastTransitionTime":"2025-11-24T17:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.646575 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.646626 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.646644 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.646666 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.646685 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:40Z","lastTransitionTime":"2025-11-24T17:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.750305 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.750371 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.750388 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.750412 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.750432 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:40Z","lastTransitionTime":"2025-11-24T17:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.852913 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.852959 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.853002 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.853019 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.853033 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:40Z","lastTransitionTime":"2025-11-24T17:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.956368 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.956425 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.956443 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.956466 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:40 crc kubenswrapper[4777]: I1124 17:01:40.956484 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:40Z","lastTransitionTime":"2025-11-24T17:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.060687 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.060855 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.060874 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.060898 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.060917 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:41Z","lastTransitionTime":"2025-11-24T17:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.163930 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.164046 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.164073 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.164105 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.164129 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:41Z","lastTransitionTime":"2025-11-24T17:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.244449 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:41 crc kubenswrapper[4777]: E1124 17:01:41.244632 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.268113 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.268184 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.268205 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.268241 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.268263 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:41Z","lastTransitionTime":"2025-11-24T17:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.372360 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.372454 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.372478 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.372511 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.372534 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:41Z","lastTransitionTime":"2025-11-24T17:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.477163 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.477263 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.477291 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.477332 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.477360 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:41Z","lastTransitionTime":"2025-11-24T17:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.582037 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.582104 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.582125 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.582159 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.582194 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:41Z","lastTransitionTime":"2025-11-24T17:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.685557 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.685631 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.685651 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.685679 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.685701 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:41Z","lastTransitionTime":"2025-11-24T17:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.789028 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.789103 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.789128 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.789162 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.789187 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:41Z","lastTransitionTime":"2025-11-24T17:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.891509 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.891568 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.891580 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.891597 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.891610 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:41Z","lastTransitionTime":"2025-11-24T17:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.962830 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.987141 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:41Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.994007 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.994153 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.994224 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.994307 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:41 crc kubenswrapper[4777]: I1124 17:01:41.994377 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:41Z","lastTransitionTime":"2025-11-24T17:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.008785 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e20e8edb-433a-4d97-b010-324307ddf3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:42Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.029361 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:42Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.049556 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:42Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.066952 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:42Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.091514 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:42Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.099255 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.099347 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.099375 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.099401 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.099420 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:42Z","lastTransitionTime":"2025-11-24T17:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.112817 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:42Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.151670 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"gressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 17:01:23.957314 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:01:23.957339 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:01:23.957286 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:01:23.957410 6297 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 17:01:23.957443 6297 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 17:01:23.957468 6297 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 17:01:23.957495 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 17:01:23.957530 6297 factory.go:656] Stopping watch factory\\\\nI1124 17:01:23.957556 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 17:01:23.957663 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 17:01:23.957688 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 17:01:23.957711 6297 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 17:01:23.957734 6297 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 17:01:23.957818 6297 ovnkube.go:599] Stopped ovnkube\\\\nI1124 17:01:23.957874 6297 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 17:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:42Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.172782 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:42Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.197812 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:42Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.207024 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.207092 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.207115 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.207158 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.207179 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:42Z","lastTransitionTime":"2025-11-24T17:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.221876 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:42Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.244798 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.244882 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.244813 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:42 crc kubenswrapper[4777]: E1124 17:01:42.245051 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:42 crc kubenswrapper[4777]: E1124 17:01:42.245167 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:42 crc kubenswrapper[4777]: E1124 17:01:42.245374 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.252154 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:42Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.275132 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:42Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.293302 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:42Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.311670 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.311744 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.311762 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.311793 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.311820 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:42Z","lastTransitionTime":"2025-11-24T17:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.313167 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:42Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.329917 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:42Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.422941 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.423040 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.423065 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.423094 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.423112 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:42Z","lastTransitionTime":"2025-11-24T17:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.526485 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.526560 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.526580 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.526608 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.526628 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:42Z","lastTransitionTime":"2025-11-24T17:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.630189 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.630271 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.630290 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.630318 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.630342 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:42Z","lastTransitionTime":"2025-11-24T17:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.734517 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.734603 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.734645 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.734674 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.734694 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:42Z","lastTransitionTime":"2025-11-24T17:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.838876 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.839480 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.839501 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.839528 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.839550 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:42Z","lastTransitionTime":"2025-11-24T17:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.943531 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.943632 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.943659 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.943692 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.943713 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:42Z","lastTransitionTime":"2025-11-24T17:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.995883 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.996159 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.996252 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:42 crc kubenswrapper[4777]: I1124 17:01:42.996296 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:42 crc kubenswrapper[4777]: E1124 17:01:42.996474 4777 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:01:42 crc kubenswrapper[4777]: E1124 17:01:42.996540 4777 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:01:42 crc kubenswrapper[4777]: E1124 17:01:42.996487 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:01:42 crc kubenswrapper[4777]: E1124 17:01:42.996642 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:01:42 crc kubenswrapper[4777]: E1124 17:01:42.996664 4777 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:42 crc kubenswrapper[4777]: E1124 17:01:42.996724 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:02:14.996690875 +0000 UTC m=+93.155885964 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:01:42 crc kubenswrapper[4777]: E1124 17:01:42.997325 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:02:14.997303434 +0000 UTC m=+93.156498513 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:42 crc kubenswrapper[4777]: E1124 17:01:42.997354 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:02:14.997339506 +0000 UTC m=+93.156534585 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:01:42 crc kubenswrapper[4777]: E1124 17:01:42.997376 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:02:14.997364976 +0000 UTC m=+93.156560055 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.049209 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.049271 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.049289 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.049314 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.049333 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:43Z","lastTransitionTime":"2025-11-24T17:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.097203 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:43 crc kubenswrapper[4777]: E1124 17:01:43.097510 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:01:43 crc kubenswrapper[4777]: E1124 17:01:43.097542 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:01:43 crc kubenswrapper[4777]: E1124 17:01:43.097566 4777 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:43 crc kubenswrapper[4777]: E1124 17:01:43.097643 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:02:15.097619375 +0000 UTC m=+93.256814464 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.152411 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.152473 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.152491 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.152517 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.152537 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:43Z","lastTransitionTime":"2025-11-24T17:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.245727 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:43 crc kubenswrapper[4777]: E1124 17:01:43.245958 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.258721 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.259306 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.259340 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.259938 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.260148 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:43Z","lastTransitionTime":"2025-11-24T17:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.273022 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.300068 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.323488 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.341148 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.360868 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.365442 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.365738 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.365914 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.366198 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.366477 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:43Z","lastTransitionTime":"2025-11-24T17:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.381190 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.406003 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.426888 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e20e8edb-433a-4d97-b010-324307ddf3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.447210 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.467059 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.470696 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.470804 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.470853 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.470879 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.470898 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:43Z","lastTransitionTime":"2025-11-24T17:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.484492 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.503579 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.524478 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.540711 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.574247 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.574287 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.574299 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.574318 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.574331 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:43Z","lastTransitionTime":"2025-11-24T17:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.578674 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"gressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 17:01:23.957314 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:01:23.957339 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:01:23.957286 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:01:23.957410 6297 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 17:01:23.957443 6297 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 17:01:23.957468 6297 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 17:01:23.957495 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 17:01:23.957530 6297 factory.go:656] Stopping watch factory\\\\nI1124 17:01:23.957556 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 17:01:23.957663 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 17:01:23.957688 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 17:01:23.957711 6297 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 17:01:23.957734 6297 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 17:01:23.957818 6297 ovnkube.go:599] Stopped ovnkube\\\\nI1124 17:01:23.957874 6297 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 17:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.594715 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.678688 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.678754 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.678772 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.678803 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.678867 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:43Z","lastTransitionTime":"2025-11-24T17:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.782411 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.782559 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.782583 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.782609 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.782629 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:43Z","lastTransitionTime":"2025-11-24T17:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.886481 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.886541 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.886560 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.886587 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.886606 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:43Z","lastTransitionTime":"2025-11-24T17:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.989123 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.989179 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.989195 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.989219 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:43 crc kubenswrapper[4777]: I1124 17:01:43.989236 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:43Z","lastTransitionTime":"2025-11-24T17:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.092655 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.092704 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.092714 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.092733 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.092745 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:44Z","lastTransitionTime":"2025-11-24T17:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.196287 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.196378 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.196393 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.196441 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.196457 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:44Z","lastTransitionTime":"2025-11-24T17:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.244939 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.244949 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.244993 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:44 crc kubenswrapper[4777]: E1124 17:01:44.245209 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:44 crc kubenswrapper[4777]: E1124 17:01:44.246063 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:44 crc kubenswrapper[4777]: E1124 17:01:44.246224 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.299488 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.299556 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.299580 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.299612 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.299635 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:44Z","lastTransitionTime":"2025-11-24T17:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.402575 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.402624 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.402639 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.402663 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.402678 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:44Z","lastTransitionTime":"2025-11-24T17:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.506593 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.506650 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.506663 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.506681 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.506695 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:44Z","lastTransitionTime":"2025-11-24T17:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.609403 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.609467 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.609496 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.609528 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.609553 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:44Z","lastTransitionTime":"2025-11-24T17:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.713865 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.713935 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.713955 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.714032 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.714058 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:44Z","lastTransitionTime":"2025-11-24T17:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.817111 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.817151 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.817165 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.817183 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.817198 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:44Z","lastTransitionTime":"2025-11-24T17:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.920037 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.920110 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.920127 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.920155 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:44 crc kubenswrapper[4777]: I1124 17:01:44.920175 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:44Z","lastTransitionTime":"2025-11-24T17:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.024007 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.024116 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.024135 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.024162 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.024180 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:45Z","lastTransitionTime":"2025-11-24T17:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.127692 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.127774 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.127795 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.127825 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.127848 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:45Z","lastTransitionTime":"2025-11-24T17:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.231114 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.231189 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.231213 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.231249 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.231272 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:45Z","lastTransitionTime":"2025-11-24T17:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.244651 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:45 crc kubenswrapper[4777]: E1124 17:01:45.244860 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.337131 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.337236 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.337259 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.337289 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.337312 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:45Z","lastTransitionTime":"2025-11-24T17:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.352756 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.352826 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.352843 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.352868 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.352888 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:45Z","lastTransitionTime":"2025-11-24T17:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:45 crc kubenswrapper[4777]: E1124 17:01:45.381100 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:45Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.386627 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.386666 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.386678 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.386698 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.386710 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:45Z","lastTransitionTime":"2025-11-24T17:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:45 crc kubenswrapper[4777]: E1124 17:01:45.407491 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:45Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.412704 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.412761 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.412790 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.412819 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.412843 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:45Z","lastTransitionTime":"2025-11-24T17:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:45 crc kubenswrapper[4777]: E1124 17:01:45.436672 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:45Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.448273 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.448323 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.448341 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.448370 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.448393 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:45Z","lastTransitionTime":"2025-11-24T17:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:45 crc kubenswrapper[4777]: E1124 17:01:45.472629 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:45Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.476817 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.476888 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.476908 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.476936 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.476958 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:45Z","lastTransitionTime":"2025-11-24T17:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:45 crc kubenswrapper[4777]: E1124 17:01:45.495639 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:45Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:45 crc kubenswrapper[4777]: E1124 17:01:45.495799 4777 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.498030 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.498084 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.498100 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.498126 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.498147 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:45Z","lastTransitionTime":"2025-11-24T17:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.601663 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.601728 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.601748 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.601779 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.601803 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:45Z","lastTransitionTime":"2025-11-24T17:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.705136 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.705202 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.705223 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.705252 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.705272 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:45Z","lastTransitionTime":"2025-11-24T17:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.808871 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.808937 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.808959 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.809019 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.809046 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:45Z","lastTransitionTime":"2025-11-24T17:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.912276 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.912348 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.912444 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.912495 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:45 crc kubenswrapper[4777]: I1124 17:01:45.912521 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:45Z","lastTransitionTime":"2025-11-24T17:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.016212 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.016285 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.016305 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.016335 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.016356 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:46Z","lastTransitionTime":"2025-11-24T17:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.119738 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.119847 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.119866 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.119898 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.119917 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:46Z","lastTransitionTime":"2025-11-24T17:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.224262 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.224319 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.224332 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.224351 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.224367 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:46Z","lastTransitionTime":"2025-11-24T17:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.244155 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.244254 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.244155 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:46 crc kubenswrapper[4777]: E1124 17:01:46.244366 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:46 crc kubenswrapper[4777]: E1124 17:01:46.244509 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:46 crc kubenswrapper[4777]: E1124 17:01:46.244741 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.327458 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.327527 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.327545 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.327570 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.327590 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:46Z","lastTransitionTime":"2025-11-24T17:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.431221 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.431288 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.431304 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.431329 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.431350 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:46Z","lastTransitionTime":"2025-11-24T17:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.534427 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.534493 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.534515 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.534547 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.534571 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:46Z","lastTransitionTime":"2025-11-24T17:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.638256 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.638324 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.638349 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.638378 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.638399 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:46Z","lastTransitionTime":"2025-11-24T17:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.742368 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.742434 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.742451 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.742480 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.742499 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:46Z","lastTransitionTime":"2025-11-24T17:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.845869 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.846332 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.846353 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.846379 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.846397 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:46Z","lastTransitionTime":"2025-11-24T17:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.949527 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.949596 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.949616 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.949642 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:46 crc kubenswrapper[4777]: I1124 17:01:46.949661 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:46Z","lastTransitionTime":"2025-11-24T17:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.052147 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.052212 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.052234 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.052266 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.052290 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:47Z","lastTransitionTime":"2025-11-24T17:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.155119 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.155201 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.155229 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.155260 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.155283 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:47Z","lastTransitionTime":"2025-11-24T17:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.245023 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:47 crc kubenswrapper[4777]: E1124 17:01:47.245299 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.258375 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.258492 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.258511 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.258539 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.258560 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:47Z","lastTransitionTime":"2025-11-24T17:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.362391 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.362528 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.362586 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.362673 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.362693 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:47Z","lastTransitionTime":"2025-11-24T17:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.466208 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.466263 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.466280 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.466306 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.466333 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:47Z","lastTransitionTime":"2025-11-24T17:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.569212 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.569287 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.569307 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.569334 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.569352 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:47Z","lastTransitionTime":"2025-11-24T17:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.672108 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.672180 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.672199 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.672227 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.672248 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:47Z","lastTransitionTime":"2025-11-24T17:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.776100 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.776153 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.776170 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.776196 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.776213 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:47Z","lastTransitionTime":"2025-11-24T17:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.879351 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.879427 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.879447 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.879475 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.879497 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:47Z","lastTransitionTime":"2025-11-24T17:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.982384 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.982455 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.982478 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.982507 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:47 crc kubenswrapper[4777]: I1124 17:01:47.982525 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:47Z","lastTransitionTime":"2025-11-24T17:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.086532 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.086596 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.086614 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.086640 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.086659 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:48Z","lastTransitionTime":"2025-11-24T17:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.190169 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.190258 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.190278 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.190307 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.190325 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:48Z","lastTransitionTime":"2025-11-24T17:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.244730 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.244806 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.244926 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:48 crc kubenswrapper[4777]: E1124 17:01:48.245146 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:48 crc kubenswrapper[4777]: E1124 17:01:48.245312 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:48 crc kubenswrapper[4777]: E1124 17:01:48.245526 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.294048 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.294114 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.294134 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.294159 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.294178 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:48Z","lastTransitionTime":"2025-11-24T17:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.398040 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.398121 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.398148 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.398208 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.398229 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:48Z","lastTransitionTime":"2025-11-24T17:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.500638 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.500691 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.500712 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.500736 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.500753 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:48Z","lastTransitionTime":"2025-11-24T17:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.604856 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.604906 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.604917 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.604941 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.604951 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:48Z","lastTransitionTime":"2025-11-24T17:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.707432 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.707471 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.707482 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.707501 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.707513 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:48Z","lastTransitionTime":"2025-11-24T17:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.810522 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.810608 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.810635 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.810666 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.810743 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:48Z","lastTransitionTime":"2025-11-24T17:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.914559 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.914617 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.914636 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.914660 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:48 crc kubenswrapper[4777]: I1124 17:01:48.914677 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:48Z","lastTransitionTime":"2025-11-24T17:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.017720 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.017771 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.017793 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.017818 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.017836 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:49Z","lastTransitionTime":"2025-11-24T17:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.120507 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.120551 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.120562 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.120579 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.120590 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:49Z","lastTransitionTime":"2025-11-24T17:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.223410 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.223475 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.223495 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.223528 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.223547 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:49Z","lastTransitionTime":"2025-11-24T17:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.249509 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:49 crc kubenswrapper[4777]: E1124 17:01:49.249659 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.250734 4777 scope.go:117] "RemoveContainer" containerID="3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.326657 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.326727 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.326746 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.326776 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.326793 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:49Z","lastTransitionTime":"2025-11-24T17:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.430712 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.430764 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.430778 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.430804 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.430820 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:49Z","lastTransitionTime":"2025-11-24T17:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.534361 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.534420 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.534431 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.534454 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.534470 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:49Z","lastTransitionTime":"2025-11-24T17:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.637245 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.637324 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.637337 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.637358 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.637372 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:49Z","lastTransitionTime":"2025-11-24T17:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.766908 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.766955 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.766981 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.767004 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.767017 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:49Z","lastTransitionTime":"2025-11-24T17:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.771649 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/1.log" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.775148 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerStarted","Data":"5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3"} Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.775761 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.794225 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:49Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.809112 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:49Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.825571 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:49Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.839244 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:49Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.850212 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:49Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.862567 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:49Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.870077 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.870117 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.870130 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.870149 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.870162 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:49Z","lastTransitionTime":"2025-11-24T17:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.874941 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:49Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.892799 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:49Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.904494 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e20e8edb-433a-4d97-b010-324307ddf3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:49Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.917591 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:49Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.931211 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:49Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.945144 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:49Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.961685 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:49Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.973022 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.973063 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.973074 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.973094 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.973106 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:49Z","lastTransitionTime":"2025-11-24T17:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.975236 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:49Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:49 crc kubenswrapper[4777]: I1124 17:01:49.996561 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"gressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 17:01:23.957314 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:01:23.957339 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:01:23.957286 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:01:23.957410 6297 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 17:01:23.957443 6297 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 17:01:23.957468 6297 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 17:01:23.957495 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 17:01:23.957530 6297 factory.go:656] Stopping watch factory\\\\nI1124 17:01:23.957556 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 17:01:23.957663 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 17:01:23.957688 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 17:01:23.957711 6297 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 17:01:23.957734 6297 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 17:01:23.957818 6297 ovnkube.go:599] Stopped ovnkube\\\\nI1124 17:01:23.957874 6297 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 17:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:49Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.007329 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:50Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.074928 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.075011 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.075024 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.075045 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.075057 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:50Z","lastTransitionTime":"2025-11-24T17:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.177534 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.177582 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.177593 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.177610 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.177623 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:50Z","lastTransitionTime":"2025-11-24T17:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.245064 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.245108 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.245136 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:50 crc kubenswrapper[4777]: E1124 17:01:50.245231 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:50 crc kubenswrapper[4777]: E1124 17:01:50.245395 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:50 crc kubenswrapper[4777]: E1124 17:01:50.245453 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.280082 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.280130 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.280143 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.280160 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.280172 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:50Z","lastTransitionTime":"2025-11-24T17:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.384830 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.385306 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.385465 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.385640 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.385776 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:50Z","lastTransitionTime":"2025-11-24T17:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.489244 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.489307 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.489327 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.489352 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.489372 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:50Z","lastTransitionTime":"2025-11-24T17:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.592392 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.592423 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.592432 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.592448 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.592458 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:50Z","lastTransitionTime":"2025-11-24T17:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.695077 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.695128 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.695140 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.695157 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.695170 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:50Z","lastTransitionTime":"2025-11-24T17:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.783159 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/2.log" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.784760 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/1.log" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.791079 4777 generic.go:334] "Generic (PLEG): container finished" podID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerID="5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3" exitCode=1 Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.791144 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerDied","Data":"5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3"} Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.791207 4777 scope.go:117] "RemoveContainer" containerID="3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.792296 4777 scope.go:117] "RemoveContainer" containerID="5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3" Nov 24 17:01:50 crc kubenswrapper[4777]: E1124 17:01:50.792700 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\"" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.806188 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.806245 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.806268 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.806302 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.806324 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:50Z","lastTransitionTime":"2025-11-24T17:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.814774 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:50Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.829949 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:50Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.850370 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c8f542662526766cd6730a8e70874ace14bca5a9cec370d61f35f31d9a6192b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:24Z\\\",\\\"message\\\":\\\"gressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 17:01:23.957314 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 17:01:23.957339 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 17:01:23.957286 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 17:01:23.957410 6297 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 17:01:23.957443 6297 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 17:01:23.957468 6297 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 17:01:23.957495 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 17:01:23.957530 6297 factory.go:656] Stopping watch factory\\\\nI1124 17:01:23.957556 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 17:01:23.957663 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 17:01:23.957688 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 17:01:23.957711 6297 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 17:01:23.957734 6297 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 17:01:23.957818 6297 ovnkube.go:599] Stopped ovnkube\\\\nI1124 17:01:23.957874 6297 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 17:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:50Z\\\",\\\"message\\\":\\\"-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1124 17:01:50.184729 6630 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1124 17:01:50.184255 6630 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1124 17:01:50.184741 6630 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1124 17:01:50.184268 6630 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:01:50.184752 6630 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1124 17:01:50.184745 6630 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nF1124 17:01:50.184760 6630 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:50Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.863304 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:50Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.879137 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:50Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.893629 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:50Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.907263 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:50Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.909721 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.909750 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.909759 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.909773 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.909784 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:50Z","lastTransitionTime":"2025-11-24T17:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.925475 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:50Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.935516 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:50Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.947838 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:50Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.958714 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:50Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.974320 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e20e8edb-433a-4d97-b010-324307ddf3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:50Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:50 crc kubenswrapper[4777]: I1124 17:01:50.996909 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:50Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.011936 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.011978 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.011988 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.012002 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.012013 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:51Z","lastTransitionTime":"2025-11-24T17:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.014108 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:51Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.027934 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:51Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.039727 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:51Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.114489 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.114553 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.114575 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.114603 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.114624 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:51Z","lastTransitionTime":"2025-11-24T17:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.217485 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.217548 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.217561 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.217577 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.217587 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:51Z","lastTransitionTime":"2025-11-24T17:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.245174 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:51 crc kubenswrapper[4777]: E1124 17:01:51.245307 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.319447 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.319479 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.319488 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.319502 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.319512 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:51Z","lastTransitionTime":"2025-11-24T17:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.422667 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.422725 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.422744 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.422774 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.422794 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:51Z","lastTransitionTime":"2025-11-24T17:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.526706 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.526770 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.526788 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.526817 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.526836 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:51Z","lastTransitionTime":"2025-11-24T17:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.630338 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.630415 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.630435 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.630464 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.630484 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:51Z","lastTransitionTime":"2025-11-24T17:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.733921 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.734018 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.734042 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.734074 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.734098 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:51Z","lastTransitionTime":"2025-11-24T17:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.798357 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/2.log" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.804726 4777 scope.go:117] "RemoveContainer" containerID="5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3" Nov 24 17:01:51 crc kubenswrapper[4777]: E1124 17:01:51.805162 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\"" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.822679 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:51Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.837090 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:51Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.837494 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.837538 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.837556 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.837580 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.837599 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:51Z","lastTransitionTime":"2025-11-24T17:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.853214 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:51Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.873233 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:50Z\\\",\\\"message\\\":\\\"-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1124 17:01:50.184729 6630 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1124 17:01:50.184255 6630 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1124 17:01:50.184741 6630 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1124 17:01:50.184268 6630 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:01:50.184752 6630 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1124 17:01:50.184745 6630 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nF1124 17:01:50.184760 6630 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:51Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.885645 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:51Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.898344 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:51Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.909706 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:51Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.925747 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:51Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.940769 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.940834 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.940857 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.940886 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.940910 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:51Z","lastTransitionTime":"2025-11-24T17:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.944303 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:51Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.963822 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:51Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.981064 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:51Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:51 crc kubenswrapper[4777]: I1124 17:01:51.995426 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:51Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.008753 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:52Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.023593 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:52Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.035180 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e20e8edb-433a-4d97-b010-324307ddf3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:52Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.044234 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.044273 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.044286 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.044304 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.044316 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:52Z","lastTransitionTime":"2025-11-24T17:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.048355 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:52Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.146585 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.146630 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.146639 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.146655 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.146669 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:52Z","lastTransitionTime":"2025-11-24T17:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.244817 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.244915 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:52 crc kubenswrapper[4777]: E1124 17:01:52.244988 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.245061 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:52 crc kubenswrapper[4777]: E1124 17:01:52.245141 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:52 crc kubenswrapper[4777]: E1124 17:01:52.245288 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.249497 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.249525 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.249533 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.249546 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.249557 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:52Z","lastTransitionTime":"2025-11-24T17:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.351963 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.352067 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.352087 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.352114 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.352138 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:52Z","lastTransitionTime":"2025-11-24T17:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.455279 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.455332 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.455342 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.455363 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.455374 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:52Z","lastTransitionTime":"2025-11-24T17:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.558265 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.558310 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.558319 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.558337 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.558348 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:52Z","lastTransitionTime":"2025-11-24T17:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.661681 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.661742 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.661761 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.661783 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.661799 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:52Z","lastTransitionTime":"2025-11-24T17:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.765873 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.765938 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.765957 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.766013 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.766034 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:52Z","lastTransitionTime":"2025-11-24T17:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.869133 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.869214 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.869235 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.869267 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.869286 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:52Z","lastTransitionTime":"2025-11-24T17:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.972156 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.972213 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.972227 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.972244 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:52 crc kubenswrapper[4777]: I1124 17:01:52.972260 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:52Z","lastTransitionTime":"2025-11-24T17:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.075082 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.075140 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.075154 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.075174 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.075185 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:53Z","lastTransitionTime":"2025-11-24T17:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.177993 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.178044 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.178054 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.178069 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.178080 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:53Z","lastTransitionTime":"2025-11-24T17:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.245020 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:53 crc kubenswrapper[4777]: E1124 17:01:53.245240 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.260896 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.280338 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.280397 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.280417 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.280441 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.280459 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:53Z","lastTransitionTime":"2025-11-24T17:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.294358 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:50Z\\\",\\\"message\\\":\\\"-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1124 17:01:50.184729 6630 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1124 17:01:50.184255 6630 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1124 17:01:50.184741 6630 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1124 17:01:50.184268 6630 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:01:50.184752 6630 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1124 17:01:50.184745 6630 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nF1124 17:01:50.184760 6630 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.310743 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.329423 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.351438 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.376427 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.382339 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.382372 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.382381 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.382396 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.382409 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:53Z","lastTransitionTime":"2025-11-24T17:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.394288 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.410278 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.430652 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.449863 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.468052 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.484689 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.484718 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.484728 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.484740 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.484750 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:53Z","lastTransitionTime":"2025-11-24T17:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.485226 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e20e8edb-433a-4d97-b010-324307ddf3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.507310 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.528107 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.548054 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.569686 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:53Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.587516 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.587564 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.587580 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.587604 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.587621 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:53Z","lastTransitionTime":"2025-11-24T17:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.690899 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.690959 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.691011 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.691047 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.691070 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:53Z","lastTransitionTime":"2025-11-24T17:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.794448 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.794495 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.794514 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.794539 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.794558 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:53Z","lastTransitionTime":"2025-11-24T17:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.898398 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.898439 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.898452 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.898468 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:53 crc kubenswrapper[4777]: I1124 17:01:53.898479 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:53Z","lastTransitionTime":"2025-11-24T17:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.001843 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.001896 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.001908 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.001925 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.001939 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:54Z","lastTransitionTime":"2025-11-24T17:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.104027 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.104108 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.104135 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.104167 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.104188 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:54Z","lastTransitionTime":"2025-11-24T17:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.206776 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.206819 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.206831 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.206850 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.206863 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:54Z","lastTransitionTime":"2025-11-24T17:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.244548 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:54 crc kubenswrapper[4777]: E1124 17:01:54.244671 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.244723 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.244792 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:54 crc kubenswrapper[4777]: E1124 17:01:54.244908 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:54 crc kubenswrapper[4777]: E1124 17:01:54.244963 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.310505 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.310563 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.310581 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.310605 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.310627 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:54Z","lastTransitionTime":"2025-11-24T17:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.414364 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.414450 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.414533 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.414572 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.414598 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:54Z","lastTransitionTime":"2025-11-24T17:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.516901 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.516961 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.517012 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.517038 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.517056 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:54Z","lastTransitionTime":"2025-11-24T17:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.620083 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.620143 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.620163 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.620189 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.620207 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:54Z","lastTransitionTime":"2025-11-24T17:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.723074 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.723123 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.723137 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.723158 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.723173 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:54Z","lastTransitionTime":"2025-11-24T17:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.735107 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs\") pod \"network-metrics-daemon-kxqzn\" (UID: \"960e1bed-abef-4f45-9475-37eee6bd780a\") " pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:54 crc kubenswrapper[4777]: E1124 17:01:54.735329 4777 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:01:54 crc kubenswrapper[4777]: E1124 17:01:54.735412 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs podName:960e1bed-abef-4f45-9475-37eee6bd780a nodeName:}" failed. No retries permitted until 2025-11-24 17:02:26.735388925 +0000 UTC m=+104.894584074 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs") pod "network-metrics-daemon-kxqzn" (UID: "960e1bed-abef-4f45-9475-37eee6bd780a") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.825719 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.825775 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.825787 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.825804 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.825819 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:54Z","lastTransitionTime":"2025-11-24T17:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.928401 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.928457 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.928473 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.928497 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:54 crc kubenswrapper[4777]: I1124 17:01:54.928516 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:54Z","lastTransitionTime":"2025-11-24T17:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.030932 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.031003 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.031016 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.031034 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.031048 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:55Z","lastTransitionTime":"2025-11-24T17:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.134090 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.134153 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.134170 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.134196 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.134214 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:55Z","lastTransitionTime":"2025-11-24T17:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.237056 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.237108 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.237122 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.237145 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.237159 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:55Z","lastTransitionTime":"2025-11-24T17:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.244787 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:55 crc kubenswrapper[4777]: E1124 17:01:55.244985 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.340393 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.340455 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.340468 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.340490 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.340772 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:55Z","lastTransitionTime":"2025-11-24T17:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.443330 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.443366 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.443377 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.443396 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.443407 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:55Z","lastTransitionTime":"2025-11-24T17:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.545727 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.545782 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.545799 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.545817 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.545830 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:55Z","lastTransitionTime":"2025-11-24T17:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.617476 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.617518 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.617528 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.617544 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.617557 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:55Z","lastTransitionTime":"2025-11-24T17:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:55 crc kubenswrapper[4777]: E1124 17:01:55.630166 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:55Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.634691 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.634741 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.634752 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.634769 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.634784 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:55Z","lastTransitionTime":"2025-11-24T17:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:55 crc kubenswrapper[4777]: E1124 17:01:55.647632 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:55Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.652778 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.653026 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.653174 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.653341 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.653461 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:55Z","lastTransitionTime":"2025-11-24T17:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:55 crc kubenswrapper[4777]: E1124 17:01:55.664417 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:55Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.668422 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.668449 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.668457 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.668474 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.668485 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:55Z","lastTransitionTime":"2025-11-24T17:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:55 crc kubenswrapper[4777]: E1124 17:01:55.681864 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:55Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.685324 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.685344 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.685352 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.685367 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.685375 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:55Z","lastTransitionTime":"2025-11-24T17:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:55 crc kubenswrapper[4777]: E1124 17:01:55.700113 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:55Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:55 crc kubenswrapper[4777]: E1124 17:01:55.700218 4777 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.701732 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.701762 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.701773 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.701786 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.701800 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:55Z","lastTransitionTime":"2025-11-24T17:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.804621 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.804667 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.804676 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.804691 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.804703 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:55Z","lastTransitionTime":"2025-11-24T17:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.908264 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.908357 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.908376 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.908405 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:55 crc kubenswrapper[4777]: I1124 17:01:55.908431 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:55Z","lastTransitionTime":"2025-11-24T17:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.011279 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.011358 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.011381 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.011410 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.011429 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:56Z","lastTransitionTime":"2025-11-24T17:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.114777 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.114830 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.114842 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.114861 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.114875 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:56Z","lastTransitionTime":"2025-11-24T17:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.217881 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.217937 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.217952 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.218002 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.218019 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:56Z","lastTransitionTime":"2025-11-24T17:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.244330 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.244424 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:56 crc kubenswrapper[4777]: E1124 17:01:56.244487 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.244511 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:56 crc kubenswrapper[4777]: E1124 17:01:56.244641 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:56 crc kubenswrapper[4777]: E1124 17:01:56.244698 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.321037 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.321127 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.321153 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.321187 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.321211 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:56Z","lastTransitionTime":"2025-11-24T17:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.423566 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.423632 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.423649 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.423673 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.423692 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:56Z","lastTransitionTime":"2025-11-24T17:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.526055 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.526097 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.526108 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.526126 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.526137 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:56Z","lastTransitionTime":"2025-11-24T17:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.629267 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.629328 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.629343 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.629370 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.629387 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:56Z","lastTransitionTime":"2025-11-24T17:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.732752 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.732816 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.732831 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.732854 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.732871 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:56Z","lastTransitionTime":"2025-11-24T17:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.836326 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.836407 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.836428 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.836454 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.836472 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:56Z","lastTransitionTime":"2025-11-24T17:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.939811 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.939879 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.939897 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.939922 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:56 crc kubenswrapper[4777]: I1124 17:01:56.939941 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:56Z","lastTransitionTime":"2025-11-24T17:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.043587 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.043662 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.043680 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.043707 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.043731 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:57Z","lastTransitionTime":"2025-11-24T17:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.146214 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.146462 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.146477 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.146497 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.146509 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:57Z","lastTransitionTime":"2025-11-24T17:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.245314 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:57 crc kubenswrapper[4777]: E1124 17:01:57.245473 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.249196 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.249231 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.249242 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.249261 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.249276 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:57Z","lastTransitionTime":"2025-11-24T17:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.352934 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.353045 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.353071 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.353100 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.353126 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:57Z","lastTransitionTime":"2025-11-24T17:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.456471 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.456538 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.456567 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.456599 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.456624 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:57Z","lastTransitionTime":"2025-11-24T17:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.559792 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.559885 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.559913 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.559947 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.559996 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:57Z","lastTransitionTime":"2025-11-24T17:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.663711 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.663788 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.663812 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.663843 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.663868 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:57Z","lastTransitionTime":"2025-11-24T17:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.766340 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.766410 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.766429 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.766456 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.766473 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:57Z","lastTransitionTime":"2025-11-24T17:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.870305 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.870377 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.870399 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.870426 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.870450 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:57Z","lastTransitionTime":"2025-11-24T17:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.973513 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.973716 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.973812 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.973895 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:57 crc kubenswrapper[4777]: I1124 17:01:57.973948 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:57Z","lastTransitionTime":"2025-11-24T17:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.077031 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.077081 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.077094 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.077119 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.077136 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:58Z","lastTransitionTime":"2025-11-24T17:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.180084 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.180505 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.180775 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.181072 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.181579 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:58Z","lastTransitionTime":"2025-11-24T17:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.244805 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.244805 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.245119 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:01:58 crc kubenswrapper[4777]: E1124 17:01:58.245348 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:01:58 crc kubenswrapper[4777]: E1124 17:01:58.245563 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:01:58 crc kubenswrapper[4777]: E1124 17:01:58.245742 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.268437 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.285023 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.285073 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.285091 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.285119 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.285138 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:58Z","lastTransitionTime":"2025-11-24T17:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.387726 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.387799 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.387822 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.387857 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.387880 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:58Z","lastTransitionTime":"2025-11-24T17:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.491010 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.491071 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.491094 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.491123 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.491148 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:58Z","lastTransitionTime":"2025-11-24T17:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.594614 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.594655 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.594672 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.594692 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.594708 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:58Z","lastTransitionTime":"2025-11-24T17:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.697066 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.697113 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.697129 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.697151 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.697166 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:58Z","lastTransitionTime":"2025-11-24T17:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.810226 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.810280 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.810292 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.810309 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.810321 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:58Z","lastTransitionTime":"2025-11-24T17:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.912217 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.912287 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.912306 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.912331 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:58 crc kubenswrapper[4777]: I1124 17:01:58.912351 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:58Z","lastTransitionTime":"2025-11-24T17:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.015404 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.015480 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.015504 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.015534 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.015557 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:59Z","lastTransitionTime":"2025-11-24T17:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.119175 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.119595 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.119750 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.119894 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.120063 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:59Z","lastTransitionTime":"2025-11-24T17:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.224236 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.224305 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.224326 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.224353 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.224374 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:59Z","lastTransitionTime":"2025-11-24T17:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.245228 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:01:59 crc kubenswrapper[4777]: E1124 17:01:59.245437 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.327755 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.327816 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.327834 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.327858 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.327880 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:59Z","lastTransitionTime":"2025-11-24T17:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.431137 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.431218 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.431237 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.431262 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.431280 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:59Z","lastTransitionTime":"2025-11-24T17:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.534309 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.534386 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.534405 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.534431 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.534453 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:59Z","lastTransitionTime":"2025-11-24T17:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.637846 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.637906 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.637923 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.637948 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.637995 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:59Z","lastTransitionTime":"2025-11-24T17:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.740855 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.740899 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.740911 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.740928 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.740941 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:59Z","lastTransitionTime":"2025-11-24T17:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.830600 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mfvtr_a5500f09-0e35-442c-87ae-8f280cd2edd1/kube-multus/0.log" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.830657 4777 generic.go:334] "Generic (PLEG): container finished" podID="a5500f09-0e35-442c-87ae-8f280cd2edd1" containerID="dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704" exitCode=1 Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.830695 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mfvtr" event={"ID":"a5500f09-0e35-442c-87ae-8f280cd2edd1","Type":"ContainerDied","Data":"dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704"} Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.831170 4777 scope.go:117] "RemoveContainer" containerID="dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.844366 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.844436 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.844456 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.844484 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.844504 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:59Z","lastTransitionTime":"2025-11-24T17:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.854712 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:59Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.877144 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:59Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.905825 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:50Z\\\",\\\"message\\\":\\\"-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1124 17:01:50.184729 6630 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1124 17:01:50.184255 6630 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1124 17:01:50.184741 6630 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1124 17:01:50.184268 6630 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:01:50.184752 6630 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1124 17:01:50.184745 6630 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nF1124 17:01:50.184760 6630 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:59Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.920309 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:59Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.930912 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bdc8331-01a9-4b35-9e2a-5d870e521fed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13ad8466a0f0b306dd124ed16530743baeb94e28b3e078e854859aaa093edf73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:59Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.947820 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:59Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.949894 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.949924 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.949960 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.950033 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.950052 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:01:59Z","lastTransitionTime":"2025-11-24T17:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.967989 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:59Z is after 2025-08-24T17:21:41Z" Nov 24 17:01:59 crc kubenswrapper[4777]: I1124 17:01:59.991034 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:01:59Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.013912 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:58Z\\\",\\\"message\\\":\\\"2025-11-24T17:01:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7\\\\n2025-11-24T17:01:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7 to /host/opt/cni/bin/\\\\n2025-11-24T17:01:13Z [verbose] multus-daemon started\\\\n2025-11-24T17:01:13Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.025927 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.040628 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.052944 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.053024 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.053037 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.053059 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.053103 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:00Z","lastTransitionTime":"2025-11-24T17:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.056034 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.079587 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.098535 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e20e8edb-433a-4d97-b010-324307ddf3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.115703 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.134719 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.154079 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.156648 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.156695 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.156708 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.156723 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.156749 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:00Z","lastTransitionTime":"2025-11-24T17:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.245279 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:00 crc kubenswrapper[4777]: E1124 17:02:00.245758 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.246223 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.246357 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:00 crc kubenswrapper[4777]: E1124 17:02:00.246544 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:00 crc kubenswrapper[4777]: E1124 17:02:00.246719 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.260064 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.260169 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.260189 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.260213 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.260233 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:00Z","lastTransitionTime":"2025-11-24T17:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.363191 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.363247 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.363269 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.363294 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.363312 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:00Z","lastTransitionTime":"2025-11-24T17:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.466470 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.466553 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.466573 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.466598 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.466617 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:00Z","lastTransitionTime":"2025-11-24T17:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.569766 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.569802 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.569814 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.569830 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.569841 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:00Z","lastTransitionTime":"2025-11-24T17:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.673045 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.673101 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.673118 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.673141 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.673160 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:00Z","lastTransitionTime":"2025-11-24T17:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.776570 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.776662 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.776680 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.776707 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.776730 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:00Z","lastTransitionTime":"2025-11-24T17:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.839373 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mfvtr_a5500f09-0e35-442c-87ae-8f280cd2edd1/kube-multus/0.log" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.839472 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mfvtr" event={"ID":"a5500f09-0e35-442c-87ae-8f280cd2edd1","Type":"ContainerStarted","Data":"899132c850ac1d7af915f5cb557c347ff8caec6f210b43743788369f5852c7e8"} Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.860816 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e20e8edb-433a-4d97-b010-324307ddf3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.880330 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.880445 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.880502 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.880529 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.880586 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:00Z","lastTransitionTime":"2025-11-24T17:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.886277 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.906214 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.926664 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.946507 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.966361 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.983780 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.983882 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.983937 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.983989 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.984009 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:00Z","lastTransitionTime":"2025-11-24T17:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:00 crc kubenswrapper[4777]: I1124 17:02:00.983905 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:00Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.015238 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:50Z\\\",\\\"message\\\":\\\"-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1124 17:01:50.184729 6630 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1124 17:01:50.184255 6630 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1124 17:01:50.184741 6630 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1124 17:01:50.184268 6630 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:01:50.184752 6630 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1124 17:01:50.184745 6630 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nF1124 17:01:50.184760 6630 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:01Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.032180 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:01Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.049766 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bdc8331-01a9-4b35-9e2a-5d870e521fed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13ad8466a0f0b306dd124ed16530743baeb94e28b3e078e854859aaa093edf73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:01Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.071566 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:01Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.087457 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.087505 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.087522 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.087549 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.087568 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:01Z","lastTransitionTime":"2025-11-24T17:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.099245 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:01Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.123422 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899132c850ac1d7af915f5cb557c347ff8caec6f210b43743788369f5852c7e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:58Z\\\",\\\"message\\\":\\\"2025-11-24T17:01:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7\\\\n2025-11-24T17:01:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7 to /host/opt/cni/bin/\\\\n2025-11-24T17:01:13Z [verbose] multus-daemon started\\\\n2025-11-24T17:01:13Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:01Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.139605 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:01Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.156210 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:01Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.172475 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:01Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.191243 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.191313 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.191331 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.191363 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.191381 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:01Z","lastTransitionTime":"2025-11-24T17:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.194097 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:01Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.245150 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:01 crc kubenswrapper[4777]: E1124 17:02:01.245394 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.295188 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.295254 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.295276 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.295302 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.295324 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:01Z","lastTransitionTime":"2025-11-24T17:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.398626 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.398693 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.398714 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.398749 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.398777 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:01Z","lastTransitionTime":"2025-11-24T17:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.502043 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.502107 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.502131 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.502161 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.502183 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:01Z","lastTransitionTime":"2025-11-24T17:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.606425 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.606502 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.606520 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.606551 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.606572 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:01Z","lastTransitionTime":"2025-11-24T17:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.716208 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.716283 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.716323 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.716360 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.716388 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:01Z","lastTransitionTime":"2025-11-24T17:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.821088 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.821168 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.821185 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.821594 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.821650 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:01Z","lastTransitionTime":"2025-11-24T17:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.925696 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.925740 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.925759 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.925782 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:01 crc kubenswrapper[4777]: I1124 17:02:01.925801 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:01Z","lastTransitionTime":"2025-11-24T17:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.029374 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.029458 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.029482 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.029515 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.029542 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:02Z","lastTransitionTime":"2025-11-24T17:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.133418 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.133478 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.133496 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.133520 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.133538 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:02Z","lastTransitionTime":"2025-11-24T17:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.238112 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.238221 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.238247 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.238283 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.238306 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:02Z","lastTransitionTime":"2025-11-24T17:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.244600 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.244727 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.244803 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:02 crc kubenswrapper[4777]: E1124 17:02:02.245124 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:02 crc kubenswrapper[4777]: E1124 17:02:02.245297 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:02 crc kubenswrapper[4777]: E1124 17:02:02.245422 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.342075 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.342153 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.342176 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.342209 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.342230 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:02Z","lastTransitionTime":"2025-11-24T17:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.446524 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.446587 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.446607 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.446633 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.446652 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:02Z","lastTransitionTime":"2025-11-24T17:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.549945 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.550093 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.550118 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.550145 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.550166 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:02Z","lastTransitionTime":"2025-11-24T17:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.653773 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.653843 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.653862 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.653889 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.653908 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:02Z","lastTransitionTime":"2025-11-24T17:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.757863 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.757931 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.757948 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.758002 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.758020 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:02Z","lastTransitionTime":"2025-11-24T17:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.862389 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.862675 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.862695 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.862726 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.862746 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:02Z","lastTransitionTime":"2025-11-24T17:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.966074 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.966205 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.966228 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.966259 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:02 crc kubenswrapper[4777]: I1124 17:02:02.966279 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:02Z","lastTransitionTime":"2025-11-24T17:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.069780 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.069859 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.069874 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.069899 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.069919 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:03Z","lastTransitionTime":"2025-11-24T17:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.172957 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.173060 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.173082 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.173112 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.173135 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:03Z","lastTransitionTime":"2025-11-24T17:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.245125 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:03 crc kubenswrapper[4777]: E1124 17:02:03.245330 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.276620 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.276705 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.276729 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.276760 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.276786 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:03Z","lastTransitionTime":"2025-11-24T17:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.278052 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:50Z\\\",\\\"message\\\":\\\"-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1124 17:01:50.184729 6630 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1124 17:01:50.184255 6630 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1124 17:01:50.184741 6630 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1124 17:01:50.184268 6630 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:01:50.184752 6630 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1124 17:01:50.184745 6630 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nF1124 17:01:50.184760 6630 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.294678 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.310049 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bdc8331-01a9-4b35-9e2a-5d870e521fed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13ad8466a0f0b306dd124ed16530743baeb94e28b3e078e854859aaa093edf73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.327599 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.344272 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.366745 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899132c850ac1d7af915f5cb557c347ff8caec6f210b43743788369f5852c7e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:58Z\\\",\\\"message\\\":\\\"2025-11-24T17:01:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7\\\\n2025-11-24T17:01:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7 to /host/opt/cni/bin/\\\\n2025-11-24T17:01:13Z [verbose] multus-daemon started\\\\n2025-11-24T17:01:13Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.384877 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.385557 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.385620 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.385644 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.385678 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.385702 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:03Z","lastTransitionTime":"2025-11-24T17:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.401760 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.420743 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.446600 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.464238 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.483100 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.488237 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.488289 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.488305 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.488329 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.488347 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:03Z","lastTransitionTime":"2025-11-24T17:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.499105 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e20e8edb-433a-4d97-b010-324307ddf3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.516386 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.538301 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.557219 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.577911 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:03Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.591368 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.591809 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.592071 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.592296 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.592453 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:03Z","lastTransitionTime":"2025-11-24T17:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.695450 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.695860 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.696057 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.696255 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.696453 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:03Z","lastTransitionTime":"2025-11-24T17:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.799908 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.800355 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.800543 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.800776 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.801015 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:03Z","lastTransitionTime":"2025-11-24T17:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.903349 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.903621 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.903683 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.903749 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:03 crc kubenswrapper[4777]: I1124 17:02:03.903815 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:03Z","lastTransitionTime":"2025-11-24T17:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.007118 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.007188 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.007210 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.007240 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.007261 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:04Z","lastTransitionTime":"2025-11-24T17:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.109759 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.109829 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.109853 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.109883 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.109909 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:04Z","lastTransitionTime":"2025-11-24T17:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.212422 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.212466 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.212485 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.212508 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.212522 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:04Z","lastTransitionTime":"2025-11-24T17:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.245305 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.245334 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.245359 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:04 crc kubenswrapper[4777]: E1124 17:02:04.245625 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:04 crc kubenswrapper[4777]: E1124 17:02:04.245752 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:04 crc kubenswrapper[4777]: E1124 17:02:04.246042 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.314864 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.314911 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.314922 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.314938 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.314949 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:04Z","lastTransitionTime":"2025-11-24T17:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.418832 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.418912 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.418930 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.418956 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.419009 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:04Z","lastTransitionTime":"2025-11-24T17:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.522216 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.522290 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.522308 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.522340 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.522360 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:04Z","lastTransitionTime":"2025-11-24T17:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.625391 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.625462 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.625482 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.625509 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.625534 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:04Z","lastTransitionTime":"2025-11-24T17:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.728347 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.728424 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.728448 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.728477 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.728499 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:04Z","lastTransitionTime":"2025-11-24T17:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.832142 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.832228 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.832285 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.832321 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.832345 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:04Z","lastTransitionTime":"2025-11-24T17:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.936238 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.936332 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.936363 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.936397 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:04 crc kubenswrapper[4777]: I1124 17:02:04.936421 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:04Z","lastTransitionTime":"2025-11-24T17:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.041403 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.041451 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.041463 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.041478 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.041488 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:05Z","lastTransitionTime":"2025-11-24T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.145036 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.145104 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.145144 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.145179 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.145205 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:05Z","lastTransitionTime":"2025-11-24T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.244866 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:05 crc kubenswrapper[4777]: E1124 17:02:05.245200 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.247964 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.248215 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.248253 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.248281 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.249143 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:05Z","lastTransitionTime":"2025-11-24T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.352958 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.353067 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.353086 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.353111 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.353129 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:05Z","lastTransitionTime":"2025-11-24T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.456510 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.456558 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.456570 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.456589 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.456601 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:05Z","lastTransitionTime":"2025-11-24T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.559767 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.560196 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.560210 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.560227 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.560237 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:05Z","lastTransitionTime":"2025-11-24T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.663997 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.664072 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.664089 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.664116 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.664134 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:05Z","lastTransitionTime":"2025-11-24T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.717031 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.717112 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.717140 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.717177 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.717203 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:05Z","lastTransitionTime":"2025-11-24T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:05 crc kubenswrapper[4777]: E1124 17:02:05.732147 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.737819 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.738069 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.738249 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.738408 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.738564 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:05Z","lastTransitionTime":"2025-11-24T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:05 crc kubenswrapper[4777]: E1124 17:02:05.763684 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.769816 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.769888 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.769917 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.769946 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.769964 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:05Z","lastTransitionTime":"2025-11-24T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:05 crc kubenswrapper[4777]: E1124 17:02:05.790305 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.796198 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.796413 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.796559 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.796690 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.796860 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:05Z","lastTransitionTime":"2025-11-24T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:05 crc kubenswrapper[4777]: E1124 17:02:05.817505 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.822478 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.822545 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.822570 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.822603 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.822629 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:05Z","lastTransitionTime":"2025-11-24T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:05 crc kubenswrapper[4777]: E1124 17:02:05.844944 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:05Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:05 crc kubenswrapper[4777]: E1124 17:02:05.845256 4777 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.847401 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.847599 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.847842 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.848093 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.848260 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:05Z","lastTransitionTime":"2025-11-24T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.950771 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.950828 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.950847 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.950869 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:05 crc kubenswrapper[4777]: I1124 17:02:05.950886 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:05Z","lastTransitionTime":"2025-11-24T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.054037 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.054395 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.054547 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.054697 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.054831 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:06Z","lastTransitionTime":"2025-11-24T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.157581 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.158401 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.158552 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.158688 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.158812 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:06Z","lastTransitionTime":"2025-11-24T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.244586 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.244670 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.244590 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:06 crc kubenswrapper[4777]: E1124 17:02:06.244898 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:06 crc kubenswrapper[4777]: E1124 17:02:06.245144 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:06 crc kubenswrapper[4777]: E1124 17:02:06.245308 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.262539 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.262602 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.262622 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.262651 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.262674 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:06Z","lastTransitionTime":"2025-11-24T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.365996 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.366065 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.366082 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.366107 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.366127 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:06Z","lastTransitionTime":"2025-11-24T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.469500 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.469576 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.469596 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.469628 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.469647 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:06Z","lastTransitionTime":"2025-11-24T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.573150 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.573590 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.574486 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.574544 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.574579 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:06Z","lastTransitionTime":"2025-11-24T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.677606 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.677688 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.677706 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.677733 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.677752 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:06Z","lastTransitionTime":"2025-11-24T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.781158 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.781208 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.781220 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.781238 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.781252 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:06Z","lastTransitionTime":"2025-11-24T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.883691 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.883777 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.883797 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.883822 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.883845 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:06Z","lastTransitionTime":"2025-11-24T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.986795 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.986868 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.986902 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.986934 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:06 crc kubenswrapper[4777]: I1124 17:02:06.986954 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:06Z","lastTransitionTime":"2025-11-24T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.090048 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.090118 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.090143 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.090175 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.090200 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:07Z","lastTransitionTime":"2025-11-24T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.193761 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.193811 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.193820 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.193838 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.193848 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:07Z","lastTransitionTime":"2025-11-24T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.246178 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:07 crc kubenswrapper[4777]: E1124 17:02:07.246384 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.247823 4777 scope.go:117] "RemoveContainer" containerID="5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3" Nov 24 17:02:07 crc kubenswrapper[4777]: E1124 17:02:07.248270 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\"" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.297163 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.297222 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.297241 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.297263 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.297283 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:07Z","lastTransitionTime":"2025-11-24T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.401207 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.401266 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.401278 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.401297 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.401313 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:07Z","lastTransitionTime":"2025-11-24T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.505375 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.505453 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.505471 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.505499 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.505526 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:07Z","lastTransitionTime":"2025-11-24T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.608815 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.608895 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.608919 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.608950 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.609025 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:07Z","lastTransitionTime":"2025-11-24T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.712452 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.712511 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.712522 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.712541 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.712554 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:07Z","lastTransitionTime":"2025-11-24T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.815081 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.815150 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.815167 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.815193 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.815212 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:07Z","lastTransitionTime":"2025-11-24T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.917315 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.917388 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.917406 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.917435 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:07 crc kubenswrapper[4777]: I1124 17:02:07.917454 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:07Z","lastTransitionTime":"2025-11-24T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.021573 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.021693 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.021713 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.021741 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.021763 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:08Z","lastTransitionTime":"2025-11-24T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.125662 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.125729 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.125745 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.125771 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.125785 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:08Z","lastTransitionTime":"2025-11-24T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.229267 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.229330 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.229347 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.229376 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.229394 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:08Z","lastTransitionTime":"2025-11-24T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.244910 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.245007 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.245089 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:08 crc kubenswrapper[4777]: E1124 17:02:08.245125 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:08 crc kubenswrapper[4777]: E1124 17:02:08.245412 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:08 crc kubenswrapper[4777]: E1124 17:02:08.246074 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.266835 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.333617 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.334185 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.334348 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.334492 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.334622 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:08Z","lastTransitionTime":"2025-11-24T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.437778 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.437871 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.437895 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.437929 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.437952 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:08Z","lastTransitionTime":"2025-11-24T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.541776 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.541893 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.541915 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.541943 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.541960 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:08Z","lastTransitionTime":"2025-11-24T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.645058 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.645126 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.645147 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.645174 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.645193 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:08Z","lastTransitionTime":"2025-11-24T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.748132 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.748195 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.748213 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.748241 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.748263 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:08Z","lastTransitionTime":"2025-11-24T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.850680 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.850722 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.850734 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.850752 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.850765 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:08Z","lastTransitionTime":"2025-11-24T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.954191 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.954285 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.954307 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.954336 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:08 crc kubenswrapper[4777]: I1124 17:02:08.954364 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:08Z","lastTransitionTime":"2025-11-24T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.057856 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.057914 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.057934 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.057959 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.058004 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:09Z","lastTransitionTime":"2025-11-24T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.162296 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.162366 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.162385 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.162410 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.162427 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:09Z","lastTransitionTime":"2025-11-24T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.245248 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:09 crc kubenswrapper[4777]: E1124 17:02:09.245466 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.266854 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.266898 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.266907 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.266921 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.266933 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:09Z","lastTransitionTime":"2025-11-24T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.373214 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.373282 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.373300 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.373323 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.373342 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:09Z","lastTransitionTime":"2025-11-24T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.476349 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.476419 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.476443 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.476469 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.476492 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:09Z","lastTransitionTime":"2025-11-24T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.579772 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.579848 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.579875 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.579910 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.579934 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:09Z","lastTransitionTime":"2025-11-24T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.684698 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.684788 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.684811 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.684842 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.684869 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:09Z","lastTransitionTime":"2025-11-24T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.788438 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.788507 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.788525 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.788554 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.788576 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:09Z","lastTransitionTime":"2025-11-24T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.892019 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.892088 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.892112 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.892142 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.892169 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:09Z","lastTransitionTime":"2025-11-24T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.995385 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.995454 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.995476 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.995507 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:09 crc kubenswrapper[4777]: I1124 17:02:09.995528 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:09Z","lastTransitionTime":"2025-11-24T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.098858 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.098936 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.098956 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.099026 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.099047 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:10Z","lastTransitionTime":"2025-11-24T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.202508 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.202637 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.202658 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.202688 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.202707 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:10Z","lastTransitionTime":"2025-11-24T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.244336 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.244401 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.244458 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:10 crc kubenswrapper[4777]: E1124 17:02:10.244535 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:10 crc kubenswrapper[4777]: E1124 17:02:10.244660 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:10 crc kubenswrapper[4777]: E1124 17:02:10.244790 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.306409 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.306475 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.306494 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.306521 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.306542 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:10Z","lastTransitionTime":"2025-11-24T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.410012 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.410067 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.410084 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.410111 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.410129 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:10Z","lastTransitionTime":"2025-11-24T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.512607 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.512671 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.512688 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.512716 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.512737 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:10Z","lastTransitionTime":"2025-11-24T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.616026 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.616113 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.616132 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.616158 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.616177 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:10Z","lastTransitionTime":"2025-11-24T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.719284 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.719350 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.719368 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.719394 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.719414 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:10Z","lastTransitionTime":"2025-11-24T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.823879 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.823947 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.823964 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.824021 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.824038 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:10Z","lastTransitionTime":"2025-11-24T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.927765 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.927823 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.927841 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.927867 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:10 crc kubenswrapper[4777]: I1124 17:02:10.928032 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:10Z","lastTransitionTime":"2025-11-24T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.031670 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.031741 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.031764 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.031793 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.031814 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:11Z","lastTransitionTime":"2025-11-24T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.134740 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.134807 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.134826 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.134855 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.134875 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:11Z","lastTransitionTime":"2025-11-24T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.239104 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.239177 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.239208 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.239242 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.239267 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:11Z","lastTransitionTime":"2025-11-24T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.244628 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:11 crc kubenswrapper[4777]: E1124 17:02:11.244827 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.342885 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.342951 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.342995 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.343020 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.343038 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:11Z","lastTransitionTime":"2025-11-24T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.446088 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.446126 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.446188 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.446205 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.446214 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:11Z","lastTransitionTime":"2025-11-24T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.549372 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.549431 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.549450 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.549475 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.549511 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:11Z","lastTransitionTime":"2025-11-24T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.653309 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.653369 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.653387 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.653413 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.653430 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:11Z","lastTransitionTime":"2025-11-24T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.757463 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.757526 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.757544 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.757574 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.757600 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:11Z","lastTransitionTime":"2025-11-24T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.861874 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.862045 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.862078 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.862140 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.862159 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:11Z","lastTransitionTime":"2025-11-24T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.965111 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.965179 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.965202 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.965247 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:11 crc kubenswrapper[4777]: I1124 17:02:11.965271 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:11Z","lastTransitionTime":"2025-11-24T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.067770 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.067825 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.067842 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.067865 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.067881 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:12Z","lastTransitionTime":"2025-11-24T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.171067 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.171138 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.171157 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.171182 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.171201 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:12Z","lastTransitionTime":"2025-11-24T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.245297 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.245355 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.245355 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:12 crc kubenswrapper[4777]: E1124 17:02:12.245513 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:12 crc kubenswrapper[4777]: E1124 17:02:12.245662 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:12 crc kubenswrapper[4777]: E1124 17:02:12.245781 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.274512 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.274567 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.274586 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.274609 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.274627 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:12Z","lastTransitionTime":"2025-11-24T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.377938 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.378056 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.378076 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.378541 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.378609 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:12Z","lastTransitionTime":"2025-11-24T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.481439 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.481879 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.481898 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.481924 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.481945 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:12Z","lastTransitionTime":"2025-11-24T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.585064 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.585117 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.585129 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.585146 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.585159 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:12Z","lastTransitionTime":"2025-11-24T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.688390 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.688453 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.688472 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.688500 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.688519 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:12Z","lastTransitionTime":"2025-11-24T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.792004 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.792050 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.792061 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.792079 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.792093 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:12Z","lastTransitionTime":"2025-11-24T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.894683 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.894767 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.894789 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.894817 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.894839 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:12Z","lastTransitionTime":"2025-11-24T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.998377 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.998439 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.998451 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.998475 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:12 crc kubenswrapper[4777]: I1124 17:02:12.998492 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:12Z","lastTransitionTime":"2025-11-24T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.101898 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.101942 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.101956 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.102014 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.102033 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:13Z","lastTransitionTime":"2025-11-24T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.205819 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.205889 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.205908 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.205934 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.205953 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:13Z","lastTransitionTime":"2025-11-24T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.244508 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:13 crc kubenswrapper[4777]: E1124 17:02:13.244706 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.268334 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e20e8edb-433a-4d97-b010-324307ddf3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.294280 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.309702 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.309811 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.309839 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.309871 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.309896 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:13Z","lastTransitionTime":"2025-11-24T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.313542 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.332354 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.353136 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.370687 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bdc8331-01a9-4b35-9e2a-5d870e521fed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13ad8466a0f0b306dd124ed16530743baeb94e28b3e078e854859aaa093edf73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.389874 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.405889 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.412564 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.412653 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.412672 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.412697 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.412719 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:13Z","lastTransitionTime":"2025-11-24T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.444887 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:50Z\\\",\\\"message\\\":\\\"-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1124 17:01:50.184729 6630 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1124 17:01:50.184255 6630 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1124 17:01:50.184741 6630 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1124 17:01:50.184268 6630 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:01:50.184752 6630 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1124 17:01:50.184745 6630 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nF1124 17:01:50.184760 6630 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.462712 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.489538 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c00fd09d-3ed4-46c9-bcff-407342b1dd40\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbf4e41a8419e135180170ee0fccae2eeb52f5ddae2c2fc6363594ec31354fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356940adde9dbdf959d83687973a3e697156f8fe126c44a8597da63bfa75275d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd39c4a740bea1727c583d538f8008b020bf44860750851fed34535ee611cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d259e984ddb8f711944a5528f33f8c7e9a9f85e6cccc5e111804a85cf9ac642f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf8026bb6460f3b94c6e148f2f6e1cec97160b9b8ac26bd022f4a828faa5d38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://104e9b5dc0c02e1ea12feeb30c7d491adc0f671b881f68a0ce4e15304554b207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://104e9b5dc0c02e1ea12feeb30c7d491adc0f671b881f68a0ce4e15304554b207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700f6f49ed9507d3c3074fc19424595e0dbc2d95b1a51a76cfa65a0b8642e68e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://700f6f49ed9507d3c3074fc19424595e0dbc2d95b1a51a76cfa65a0b8642e68e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://55297085683d62c3b6c607487741e50eb2ac526c3304b140384bb33f85ce6d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55297085683d62c3b6c607487741e50eb2ac526c3304b140384bb33f85ce6d6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.509566 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.515918 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.516013 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.516033 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.516068 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.516088 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:13Z","lastTransitionTime":"2025-11-24T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.529732 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.556079 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.578943 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899132c850ac1d7af915f5cb557c347ff8caec6f210b43743788369f5852c7e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:58Z\\\",\\\"message\\\":\\\"2025-11-24T17:01:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7\\\\n2025-11-24T17:01:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7 to /host/opt/cni/bin/\\\\n2025-11-24T17:01:13Z [verbose] multus-daemon started\\\\n2025-11-24T17:01:13Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.600238 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.615408 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.620088 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.620141 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.620166 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.620198 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.620223 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:13Z","lastTransitionTime":"2025-11-24T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.632749 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:13Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.723733 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.723795 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.723818 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.723851 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.723873 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:13Z","lastTransitionTime":"2025-11-24T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.826574 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.826639 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.826662 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.826693 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.826715 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:13Z","lastTransitionTime":"2025-11-24T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.930242 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.930290 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.930309 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.930333 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:13 crc kubenswrapper[4777]: I1124 17:02:13.930353 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:13Z","lastTransitionTime":"2025-11-24T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.037608 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.037697 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.037726 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.037763 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.037789 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:14Z","lastTransitionTime":"2025-11-24T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.141835 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.142327 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.142489 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.142641 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.142776 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:14Z","lastTransitionTime":"2025-11-24T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.244390 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.244468 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:14 crc kubenswrapper[4777]: E1124 17:02:14.244769 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.244489 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:14 crc kubenswrapper[4777]: E1124 17:02:14.245385 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:14 crc kubenswrapper[4777]: E1124 17:02:14.245062 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.245615 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.245904 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.246135 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.246298 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.246430 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:14Z","lastTransitionTime":"2025-11-24T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.349869 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.349950 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.350016 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.350049 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.350072 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:14Z","lastTransitionTime":"2025-11-24T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.452639 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.452719 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.452740 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.452773 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.452796 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:14Z","lastTransitionTime":"2025-11-24T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.556509 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.556642 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.556655 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.556680 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.556694 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:14Z","lastTransitionTime":"2025-11-24T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.659875 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.659942 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.659963 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.660023 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.660042 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:14Z","lastTransitionTime":"2025-11-24T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.763116 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.763532 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.763692 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.763869 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.764061 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:14Z","lastTransitionTime":"2025-11-24T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.868140 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.868197 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.868210 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.868229 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.868242 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:14Z","lastTransitionTime":"2025-11-24T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.971407 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.971477 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.971495 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.971521 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:14 crc kubenswrapper[4777]: I1124 17:02:14.971540 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:14Z","lastTransitionTime":"2025-11-24T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.074497 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.074569 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.074586 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.074613 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.074632 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:15Z","lastTransitionTime":"2025-11-24T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.083371 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.083595 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:15 crc kubenswrapper[4777]: E1124 17:02:15.083616 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:19.083579161 +0000 UTC m=+157.242774250 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:02:15 crc kubenswrapper[4777]: E1124 17:02:15.083716 4777 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.083746 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:15 crc kubenswrapper[4777]: E1124 17:02:15.083781 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:03:19.083765257 +0000 UTC m=+157.242960346 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.083888 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:15 crc kubenswrapper[4777]: E1124 17:02:15.083945 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:02:15 crc kubenswrapper[4777]: E1124 17:02:15.084005 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:02:15 crc kubenswrapper[4777]: E1124 17:02:15.084028 4777 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:02:15 crc kubenswrapper[4777]: E1124 17:02:15.084040 4777 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:02:15 crc kubenswrapper[4777]: E1124 17:02:15.084095 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:03:19.084075977 +0000 UTC m=+157.243271066 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:02:15 crc kubenswrapper[4777]: E1124 17:02:15.084122 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:03:19.084109898 +0000 UTC m=+157.243304987 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.177897 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.178075 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.178114 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.179933 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.180052 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:15Z","lastTransitionTime":"2025-11-24T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.185365 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:15 crc kubenswrapper[4777]: E1124 17:02:15.185880 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:02:15 crc kubenswrapper[4777]: E1124 17:02:15.186046 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:02:15 crc kubenswrapper[4777]: E1124 17:02:15.186076 4777 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:02:15 crc kubenswrapper[4777]: E1124 17:02:15.186241 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:03:19.186158957 +0000 UTC m=+157.345354036 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.245276 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:15 crc kubenswrapper[4777]: E1124 17:02:15.245762 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.283144 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.283362 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.283385 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.283411 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.283435 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:15Z","lastTransitionTime":"2025-11-24T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.385928 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.386056 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.386112 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.386139 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.386159 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:15Z","lastTransitionTime":"2025-11-24T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.500955 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.501070 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.501095 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.501183 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.501213 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:15Z","lastTransitionTime":"2025-11-24T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.604241 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.604342 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.604371 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.604401 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.604421 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:15Z","lastTransitionTime":"2025-11-24T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.708300 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.708379 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.708397 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.708425 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.708444 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:15Z","lastTransitionTime":"2025-11-24T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.811313 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.811386 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.811403 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.811431 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.811450 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:15Z","lastTransitionTime":"2025-11-24T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.915019 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.915084 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.915103 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.915128 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.915152 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:15Z","lastTransitionTime":"2025-11-24T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.995296 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.995695 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.995868 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.996041 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:15 crc kubenswrapper[4777]: I1124 17:02:15.996206 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:15Z","lastTransitionTime":"2025-11-24T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:16 crc kubenswrapper[4777]: E1124 17:02:16.017625 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.023951 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.024049 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.024070 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.024097 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.024119 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:16Z","lastTransitionTime":"2025-11-24T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:16 crc kubenswrapper[4777]: E1124 17:02:16.046133 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.052021 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.052091 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.052121 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.052154 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.052179 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:16Z","lastTransitionTime":"2025-11-24T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:16 crc kubenswrapper[4777]: E1124 17:02:16.074017 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.080412 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.080493 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.080521 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.080551 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.080573 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:16Z","lastTransitionTime":"2025-11-24T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:16 crc kubenswrapper[4777]: E1124 17:02:16.100941 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.106662 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.106884 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.107098 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.107262 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.107388 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:16Z","lastTransitionTime":"2025-11-24T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:16 crc kubenswrapper[4777]: E1124 17:02:16.128024 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:16Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:16 crc kubenswrapper[4777]: E1124 17:02:16.128616 4777 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.130940 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.131054 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.131079 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.131108 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.131130 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:16Z","lastTransitionTime":"2025-11-24T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.234543 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.234624 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.234644 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.234685 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.234711 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:16Z","lastTransitionTime":"2025-11-24T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.245018 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.245048 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.245086 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:16 crc kubenswrapper[4777]: E1124 17:02:16.245232 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:16 crc kubenswrapper[4777]: E1124 17:02:16.245375 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:16 crc kubenswrapper[4777]: E1124 17:02:16.245539 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.338406 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.338477 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.338499 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.338526 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.338548 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:16Z","lastTransitionTime":"2025-11-24T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.442457 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.442553 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.442580 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.442610 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.442632 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:16Z","lastTransitionTime":"2025-11-24T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.545029 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.545223 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.545271 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.545310 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.545334 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:16Z","lastTransitionTime":"2025-11-24T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.648788 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.648868 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.648904 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.648937 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.648962 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:16Z","lastTransitionTime":"2025-11-24T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.752556 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.752656 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.752684 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.752726 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.752753 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:16Z","lastTransitionTime":"2025-11-24T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.856853 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.857012 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.857041 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.857073 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.857095 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:16Z","lastTransitionTime":"2025-11-24T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.960942 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.961034 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.961055 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.961081 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:16 crc kubenswrapper[4777]: I1124 17:02:16.961099 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:16Z","lastTransitionTime":"2025-11-24T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.064560 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.064645 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.064663 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.064690 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.064711 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:17Z","lastTransitionTime":"2025-11-24T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.167874 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.167940 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.168013 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.168048 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.168070 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:17Z","lastTransitionTime":"2025-11-24T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.245066 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:17 crc kubenswrapper[4777]: E1124 17:02:17.245296 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.271545 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.271611 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.271630 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.271653 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.271672 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:17Z","lastTransitionTime":"2025-11-24T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.375174 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.375257 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.375276 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.375305 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.375325 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:17Z","lastTransitionTime":"2025-11-24T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.484726 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.484882 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.484905 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.484934 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.484953 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:17Z","lastTransitionTime":"2025-11-24T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.588378 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.589152 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.589182 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.589208 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.589226 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:17Z","lastTransitionTime":"2025-11-24T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.695359 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.695413 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.695425 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.695443 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.695455 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:17Z","lastTransitionTime":"2025-11-24T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.799600 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.799662 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.799679 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.799703 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.799721 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:17Z","lastTransitionTime":"2025-11-24T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.903606 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.903665 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.903687 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.903713 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:17 crc kubenswrapper[4777]: I1124 17:02:17.903733 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:17Z","lastTransitionTime":"2025-11-24T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.007129 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.007242 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.007261 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.007311 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.007330 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:18Z","lastTransitionTime":"2025-11-24T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.112797 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.112864 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.112883 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.112909 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.112928 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:18Z","lastTransitionTime":"2025-11-24T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.216511 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.216885 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.217045 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.217420 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.217575 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:18Z","lastTransitionTime":"2025-11-24T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.244478 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.244518 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.244523 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:18 crc kubenswrapper[4777]: E1124 17:02:18.244680 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:18 crc kubenswrapper[4777]: E1124 17:02:18.244940 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:18 crc kubenswrapper[4777]: E1124 17:02:18.245037 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.320909 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.321145 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.321369 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.321475 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.321566 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:18Z","lastTransitionTime":"2025-11-24T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.425132 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.425551 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.425735 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.425917 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.426157 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:18Z","lastTransitionTime":"2025-11-24T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.529730 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.529810 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.529830 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.529860 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.529881 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:18Z","lastTransitionTime":"2025-11-24T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.633462 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.633547 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.633567 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.633602 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.633623 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:18Z","lastTransitionTime":"2025-11-24T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.737665 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.737746 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.737764 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.737793 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.737818 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:18Z","lastTransitionTime":"2025-11-24T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.841538 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.841605 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.841628 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.841658 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.841680 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:18Z","lastTransitionTime":"2025-11-24T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.944389 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.944520 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.944547 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.944581 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:18 crc kubenswrapper[4777]: I1124 17:02:18.944604 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:18Z","lastTransitionTime":"2025-11-24T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.048262 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.048358 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.048383 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.048411 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.048430 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:19Z","lastTransitionTime":"2025-11-24T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.152235 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.152311 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.152331 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.152363 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.152381 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:19Z","lastTransitionTime":"2025-11-24T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.245066 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:19 crc kubenswrapper[4777]: E1124 17:02:19.245310 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.246664 4777 scope.go:117] "RemoveContainer" containerID="5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.255215 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.255270 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.255294 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.255318 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.255338 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:19Z","lastTransitionTime":"2025-11-24T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.359225 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.359633 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.359653 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.359679 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.359700 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:19Z","lastTransitionTime":"2025-11-24T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.462847 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.462915 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.462938 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.463000 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.463028 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:19Z","lastTransitionTime":"2025-11-24T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.566772 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.566823 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.566837 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.566856 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.566870 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:19Z","lastTransitionTime":"2025-11-24T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.670191 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.670272 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.670298 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.670336 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.670378 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:19Z","lastTransitionTime":"2025-11-24T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.778454 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.778505 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.778520 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.778542 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.778555 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:19Z","lastTransitionTime":"2025-11-24T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.881108 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.881164 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.881178 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.881198 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.881213 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:19Z","lastTransitionTime":"2025-11-24T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.914123 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/2.log" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.917160 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerStarted","Data":"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e"} Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.917908 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.931087 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.949363 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.975049 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c00fd09d-3ed4-46c9-bcff-407342b1dd40\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbf4e41a8419e135180170ee0fccae2eeb52f5ddae2c2fc6363594ec31354fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356940adde9dbdf959d83687973a3e697156f8fe126c44a8597da63bfa75275d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd39c4a740bea1727c583d538f8008b020bf44860750851fed34535ee611cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d259e984ddb8f711944a5528f33f8c7e9a9f85e6cccc5e111804a85cf9ac642f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf8026bb6460f3b94c6e148f2f6e1cec97160b9b8ac26bd022f4a828faa5d38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://104e9b5dc0c02e1ea12feeb30c7d491adc0f671b881f68a0ce4e15304554b207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://104e9b5dc0c02e1ea12feeb30c7d491adc0f671b881f68a0ce4e15304554b207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700f6f49ed9507d3c3074fc19424595e0dbc2d95b1a51a76cfa65a0b8642e68e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://700f6f49ed9507d3c3074fc19424595e0dbc2d95b1a51a76cfa65a0b8642e68e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://55297085683d62c3b6c607487741e50eb2ac526c3304b140384bb33f85ce6d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55297085683d62c3b6c607487741e50eb2ac526c3304b140384bb33f85ce6d6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.983755 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.983831 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.983854 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.983888 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.983914 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:19Z","lastTransitionTime":"2025-11-24T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:19 crc kubenswrapper[4777]: I1124 17:02:19.991670 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:19Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.011037 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.034010 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.051294 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899132c850ac1d7af915f5cb557c347ff8caec6f210b43743788369f5852c7e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:58Z\\\",\\\"message\\\":\\\"2025-11-24T17:01:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7\\\\n2025-11-24T17:01:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7 to /host/opt/cni/bin/\\\\n2025-11-24T17:01:13Z [verbose] multus-daemon started\\\\n2025-11-24T17:01:13Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.062824 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.077737 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e20e8edb-433a-4d97-b010-324307ddf3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.086406 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.086464 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.086477 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.086496 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.086510 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:20Z","lastTransitionTime":"2025-11-24T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.095105 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.112130 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.128333 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.147287 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.162465 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bdc8331-01a9-4b35-9e2a-5d870e521fed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13ad8466a0f0b306dd124ed16530743baeb94e28b3e078e854859aaa093edf73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.179606 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.190150 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.190236 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.190259 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.190288 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.190309 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:20Z","lastTransitionTime":"2025-11-24T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.197243 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.221038 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:50Z\\\",\\\"message\\\":\\\"-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1124 17:01:50.184729 6630 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1124 17:01:50.184255 6630 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1124 17:01:50.184741 6630 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1124 17:01:50.184268 6630 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:01:50.184752 6630 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1124 17:01:50.184745 6630 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nF1124 17:01:50.184760 6630 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.233656 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.245153 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.245206 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.245346 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:20 crc kubenswrapper[4777]: E1124 17:02:20.245523 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:20 crc kubenswrapper[4777]: E1124 17:02:20.245809 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:20 crc kubenswrapper[4777]: E1124 17:02:20.245874 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.294133 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.294184 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.294226 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.294247 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.294260 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:20Z","lastTransitionTime":"2025-11-24T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.397713 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.397785 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.397805 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.397833 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.397853 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:20Z","lastTransitionTime":"2025-11-24T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.501014 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.501085 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.501105 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.501132 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.501150 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:20Z","lastTransitionTime":"2025-11-24T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.604723 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.604767 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.604777 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.604794 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.604807 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:20Z","lastTransitionTime":"2025-11-24T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.708920 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.710080 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.710196 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.710342 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.710553 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:20Z","lastTransitionTime":"2025-11-24T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.813918 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.814027 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.814050 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.814074 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.814091 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:20Z","lastTransitionTime":"2025-11-24T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.917171 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.917248 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.917271 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.917304 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.917419 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:20Z","lastTransitionTime":"2025-11-24T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.923774 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/3.log" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.924627 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/2.log" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.928824 4777 generic.go:334] "Generic (PLEG): container finished" podID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerID="1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e" exitCode=1 Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.928886 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerDied","Data":"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e"} Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.928938 4777 scope.go:117] "RemoveContainer" containerID="5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.930246 4777 scope.go:117] "RemoveContainer" containerID="1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e" Nov 24 17:02:20 crc kubenswrapper[4777]: E1124 17:02:20.930630 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\"" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" Nov 24 17:02:20 crc kubenswrapper[4777]: I1124 17:02:20.966073 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:20Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.004815 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c00fd09d-3ed4-46c9-bcff-407342b1dd40\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbf4e41a8419e135180170ee0fccae2eeb52f5ddae2c2fc6363594ec31354fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356940adde9dbdf959d83687973a3e697156f8fe126c44a8597da63bfa75275d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd39c4a740bea1727c583d538f8008b020bf44860750851fed34535ee611cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d259e984ddb8f711944a5528f33f8c7e9a9f85e6cccc5e111804a85cf9ac642f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf8026bb6460f3b94c6e148f2f6e1cec97160b9b8ac26bd022f4a828faa5d38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://104e9b5dc0c02e1ea12feeb30c7d491adc0f671b881f68a0ce4e15304554b207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://104e9b5dc0c02e1ea12feeb30c7d491adc0f671b881f68a0ce4e15304554b207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700f6f49ed9507d3c3074fc19424595e0dbc2d95b1a51a76cfa65a0b8642e68e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://700f6f49ed9507d3c3074fc19424595e0dbc2d95b1a51a76cfa65a0b8642e68e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://55297085683d62c3b6c607487741e50eb2ac526c3304b140384bb33f85ce6d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55297085683d62c3b6c607487741e50eb2ac526c3304b140384bb33f85ce6d6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.020437 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.020485 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.020496 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.020514 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.020526 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:21Z","lastTransitionTime":"2025-11-24T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.028215 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.050831 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.078011 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.097598 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899132c850ac1d7af915f5cb557c347ff8caec6f210b43743788369f5852c7e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:58Z\\\",\\\"message\\\":\\\"2025-11-24T17:01:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7\\\\n2025-11-24T17:01:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7 to /host/opt/cni/bin/\\\\n2025-11-24T17:01:13Z [verbose] multus-daemon started\\\\n2025-11-24T17:01:13Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.116536 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.123853 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.123931 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.123962 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.124046 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.124071 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:21Z","lastTransitionTime":"2025-11-24T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.140390 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.162958 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e20e8edb-433a-4d97-b010-324307ddf3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.185437 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.207445 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.226943 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.227215 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.227283 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.227296 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.227341 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.227384 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:21Z","lastTransitionTime":"2025-11-24T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.244717 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:21 crc kubenswrapper[4777]: E1124 17:02:21.244948 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.248801 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.266724 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bdc8331-01a9-4b35-9e2a-5d870e521fed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13ad8466a0f0b306dd124ed16530743baeb94e28b3e078e854859aaa093edf73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.288663 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.317636 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.330758 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.330837 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.330859 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.330890 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.330910 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:21Z","lastTransitionTime":"2025-11-24T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.360825 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cfd5f4014f0fb0035573a1eb9c6ab509820d32f2c744a2ee8cbeea0c46cc4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:50Z\\\",\\\"message\\\":\\\"-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1124 17:01:50.184729 6630 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1124 17:01:50.184255 6630 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1124 17:01:50.184741 6630 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1124 17:01:50.184268 6630 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1124 17:01:50.184752 6630 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1124 17:01:50.184745 6630 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nF1124 17:01:50.184760 6630 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:02:20Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-etcd/etcd\\\\\\\"}\\\\nI1124 17:02:20.267327 6959 services_controller.go:360] Finished syncing service etcd on namespace openshift-etcd for network=default : 1.404495ms\\\\nI1124 17:02:20.267342 6959 services_controller.go:356] Processing sync for service openshift-machine-api/cluster-autoscaler-operator for network=default\\\\nI1124 17:02:20.267160 6959 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1124 17:02:20.267364 6959 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 17:02:20.267426 6959 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.382116 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.434611 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.434686 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.434705 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.434734 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.434762 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:21Z","lastTransitionTime":"2025-11-24T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.538644 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.538704 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.538724 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.538751 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.538772 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:21Z","lastTransitionTime":"2025-11-24T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.643313 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.643402 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.643421 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.643452 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.643472 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:21Z","lastTransitionTime":"2025-11-24T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.746565 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.746648 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.746671 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.746706 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.746732 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:21Z","lastTransitionTime":"2025-11-24T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.850568 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.850623 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.850637 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.850659 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.850674 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:21Z","lastTransitionTime":"2025-11-24T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.935541 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/3.log" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.941508 4777 scope.go:117] "RemoveContainer" containerID="1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e" Nov 24 17:02:21 crc kubenswrapper[4777]: E1124 17:02:21.941939 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\"" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.953896 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.953943 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.953957 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.954003 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.954021 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:21Z","lastTransitionTime":"2025-11-24T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.961874 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e20e8edb-433a-4d97-b010-324307ddf3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:21 crc kubenswrapper[4777]: I1124 17:02:21.985304 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:21Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.006507 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.025384 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.044157 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.057941 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.058027 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.058047 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.058077 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.058103 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:22Z","lastTransitionTime":"2025-11-24T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.069535 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bdc8331-01a9-4b35-9e2a-5d870e521fed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13ad8466a0f0b306dd124ed16530743baeb94e28b3e078e854859aaa093edf73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.091549 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.108505 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.131119 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:02:20Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-etcd/etcd\\\\\\\"}\\\\nI1124 17:02:20.267327 6959 services_controller.go:360] Finished syncing service etcd on namespace openshift-etcd for network=default : 1.404495ms\\\\nI1124 17:02:20.267342 6959 services_controller.go:356] Processing sync for service openshift-machine-api/cluster-autoscaler-operator for network=default\\\\nI1124 17:02:20.267160 6959 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1124 17:02:20.267364 6959 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 17:02:20.267426 6959 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:02:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.151736 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.162260 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.162328 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.162349 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.162378 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.162399 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:22Z","lastTransitionTime":"2025-11-24T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.182743 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c00fd09d-3ed4-46c9-bcff-407342b1dd40\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbf4e41a8419e135180170ee0fccae2eeb52f5ddae2c2fc6363594ec31354fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356940adde9dbdf959d83687973a3e697156f8fe126c44a8597da63bfa75275d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd39c4a740bea1727c583d538f8008b020bf44860750851fed34535ee611cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d259e984ddb8f711944a5528f33f8c7e9a9f85e6cccc5e111804a85cf9ac642f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf8026bb6460f3b94c6e148f2f6e1cec97160b9b8ac26bd022f4a828faa5d38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://104e9b5dc0c02e1ea12feeb30c7d491adc0f671b881f68a0ce4e15304554b207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://104e9b5dc0c02e1ea12feeb30c7d491adc0f671b881f68a0ce4e15304554b207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700f6f49ed9507d3c3074fc19424595e0dbc2d95b1a51a76cfa65a0b8642e68e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://700f6f49ed9507d3c3074fc19424595e0dbc2d95b1a51a76cfa65a0b8642e68e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://55297085683d62c3b6c607487741e50eb2ac526c3304b140384bb33f85ce6d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55297085683d62c3b6c607487741e50eb2ac526c3304b140384bb33f85ce6d6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.203320 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.221581 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.242577 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.244742 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:22 crc kubenswrapper[4777]: E1124 17:02:22.244938 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.245141 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.245144 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:22 crc kubenswrapper[4777]: E1124 17:02:22.245393 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:22 crc kubenswrapper[4777]: E1124 17:02:22.245530 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.266351 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899132c850ac1d7af915f5cb557c347ff8caec6f210b43743788369f5852c7e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:58Z\\\",\\\"message\\\":\\\"2025-11-24T17:01:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7\\\\n2025-11-24T17:01:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7 to /host/opt/cni/bin/\\\\n2025-11-24T17:01:13Z [verbose] multus-daemon started\\\\n2025-11-24T17:01:13Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.266399 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.266583 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.266596 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.266615 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.266629 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:22Z","lastTransitionTime":"2025-11-24T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.285250 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.306003 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.326922 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:22Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.371148 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.371207 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.371219 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.371241 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.371256 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:22Z","lastTransitionTime":"2025-11-24T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.474577 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.474642 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.474662 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.474690 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.474711 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:22Z","lastTransitionTime":"2025-11-24T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.577900 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.578012 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.578041 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.578071 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.578090 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:22Z","lastTransitionTime":"2025-11-24T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.681043 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.681134 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.681152 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.681175 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.681193 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:22Z","lastTransitionTime":"2025-11-24T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.784377 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.784425 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.784440 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.784459 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.784472 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:22Z","lastTransitionTime":"2025-11-24T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.887169 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.887234 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.887261 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.887293 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.887313 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:22Z","lastTransitionTime":"2025-11-24T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.991135 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.991201 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.991220 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.991248 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:22 crc kubenswrapper[4777]: I1124 17:02:22.991268 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:22Z","lastTransitionTime":"2025-11-24T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.093919 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.093998 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.094017 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.094043 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.094087 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:23Z","lastTransitionTime":"2025-11-24T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.197487 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.197558 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.197698 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.197742 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.197766 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:23Z","lastTransitionTime":"2025-11-24T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.245179 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:23 crc kubenswrapper[4777]: E1124 17:02:23.245573 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.267650 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.288300 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.301137 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.301186 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.301202 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.301226 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.301245 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:23Z","lastTransitionTime":"2025-11-24T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.315891 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.338489 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899132c850ac1d7af915f5cb557c347ff8caec6f210b43743788369f5852c7e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:58Z\\\",\\\"message\\\":\\\"2025-11-24T17:01:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7\\\\n2025-11-24T17:01:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7 to /host/opt/cni/bin/\\\\n2025-11-24T17:01:13Z [verbose] multus-daemon started\\\\n2025-11-24T17:01:13Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.356560 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.376271 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.393874 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.405438 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.405494 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.405510 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.405536 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.405554 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:23Z","lastTransitionTime":"2025-11-24T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.428453 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c00fd09d-3ed4-46c9-bcff-407342b1dd40\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbf4e41a8419e135180170ee0fccae2eeb52f5ddae2c2fc6363594ec31354fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356940adde9dbdf959d83687973a3e697156f8fe126c44a8597da63bfa75275d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd39c4a740bea1727c583d538f8008b020bf44860750851fed34535ee611cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d259e984ddb8f711944a5528f33f8c7e9a9f85e6cccc5e111804a85cf9ac642f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf8026bb6460f3b94c6e148f2f6e1cec97160b9b8ac26bd022f4a828faa5d38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://104e9b5dc0c02e1ea12feeb30c7d491adc0f671b881f68a0ce4e15304554b207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://104e9b5dc0c02e1ea12feeb30c7d491adc0f671b881f68a0ce4e15304554b207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700f6f49ed9507d3c3074fc19424595e0dbc2d95b1a51a76cfa65a0b8642e68e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://700f6f49ed9507d3c3074fc19424595e0dbc2d95b1a51a76cfa65a0b8642e68e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://55297085683d62c3b6c607487741e50eb2ac526c3304b140384bb33f85ce6d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55297085683d62c3b6c607487741e50eb2ac526c3304b140384bb33f85ce6d6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.451659 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.470273 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e20e8edb-433a-4d97-b010-324307ddf3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.488648 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.508829 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.508875 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.508887 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.508903 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.508914 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:23Z","lastTransitionTime":"2025-11-24T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.513062 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.538858 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.560608 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.582454 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.617939 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.618055 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.618077 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.618106 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.618159 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:23Z","lastTransitionTime":"2025-11-24T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.618674 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:02:20Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-etcd/etcd\\\\\\\"}\\\\nI1124 17:02:20.267327 6959 services_controller.go:360] Finished syncing service etcd on namespace openshift-etcd for network=default : 1.404495ms\\\\nI1124 17:02:20.267342 6959 services_controller.go:356] Processing sync for service openshift-machine-api/cluster-autoscaler-operator for network=default\\\\nI1124 17:02:20.267160 6959 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1124 17:02:20.267364 6959 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 17:02:20.267426 6959 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:02:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.637503 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.654138 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bdc8331-01a9-4b35-9e2a-5d870e521fed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13ad8466a0f0b306dd124ed16530743baeb94e28b3e078e854859aaa093edf73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:23Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.721908 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.722009 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.722034 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.722064 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.722083 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:23Z","lastTransitionTime":"2025-11-24T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.825219 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.825292 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.825312 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.825339 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.825359 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:23Z","lastTransitionTime":"2025-11-24T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.929448 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.929554 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.929579 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.929613 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:23 crc kubenswrapper[4777]: I1124 17:02:23.929633 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:23Z","lastTransitionTime":"2025-11-24T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.033581 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.033651 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.033664 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.033685 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.033698 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:24Z","lastTransitionTime":"2025-11-24T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.137397 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.137556 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.137583 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.137616 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.137648 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:24Z","lastTransitionTime":"2025-11-24T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.241666 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.241732 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.241752 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.241779 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.241802 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:24Z","lastTransitionTime":"2025-11-24T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.245118 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.245164 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.245227 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:24 crc kubenswrapper[4777]: E1124 17:02:24.245312 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:24 crc kubenswrapper[4777]: E1124 17:02:24.245452 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:24 crc kubenswrapper[4777]: E1124 17:02:24.245671 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.344380 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.344443 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.344464 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.344496 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.344518 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:24Z","lastTransitionTime":"2025-11-24T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.448459 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.448522 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.448548 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.448578 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.448603 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:24Z","lastTransitionTime":"2025-11-24T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.551311 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.551392 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.551412 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.551441 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.551461 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:24Z","lastTransitionTime":"2025-11-24T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.654507 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.654588 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.654604 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.654629 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.654646 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:24Z","lastTransitionTime":"2025-11-24T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.758358 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.758415 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.758427 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.758445 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.758457 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:24Z","lastTransitionTime":"2025-11-24T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.862847 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.862892 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.862902 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.862918 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.862931 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:24Z","lastTransitionTime":"2025-11-24T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.966841 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.966916 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.966935 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.966963 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:24 crc kubenswrapper[4777]: I1124 17:02:24.967006 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:24Z","lastTransitionTime":"2025-11-24T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.081653 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.081722 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.081741 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.081767 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.081789 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:25Z","lastTransitionTime":"2025-11-24T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.185751 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.185805 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.185817 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.185840 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.185854 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:25Z","lastTransitionTime":"2025-11-24T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.244500 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:25 crc kubenswrapper[4777]: E1124 17:02:25.244920 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.289309 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.289360 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.289375 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.289396 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.289411 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:25Z","lastTransitionTime":"2025-11-24T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.392403 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.392473 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.392497 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.392525 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.392545 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:25Z","lastTransitionTime":"2025-11-24T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.496430 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.496491 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.496509 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.496534 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.496554 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:25Z","lastTransitionTime":"2025-11-24T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.600042 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.600118 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.600137 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.600163 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.600183 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:25Z","lastTransitionTime":"2025-11-24T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.704325 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.704512 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.704534 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.704566 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.704586 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:25Z","lastTransitionTime":"2025-11-24T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.808011 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.808067 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.808081 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.808103 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.808118 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:25Z","lastTransitionTime":"2025-11-24T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.911016 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.911076 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.911091 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.911115 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:25 crc kubenswrapper[4777]: I1124 17:02:25.911134 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:25Z","lastTransitionTime":"2025-11-24T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.014215 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.014279 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.014298 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.014324 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.014346 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:26Z","lastTransitionTime":"2025-11-24T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.118162 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.118248 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.118280 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.118313 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.118337 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:26Z","lastTransitionTime":"2025-11-24T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.221544 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.221610 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.221629 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.221655 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.221673 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:26Z","lastTransitionTime":"2025-11-24T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.244523 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.244576 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.244596 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:26 crc kubenswrapper[4777]: E1124 17:02:26.244736 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:26 crc kubenswrapper[4777]: E1124 17:02:26.244901 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:26 crc kubenswrapper[4777]: E1124 17:02:26.245093 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.325472 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.325539 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.325559 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.325587 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.325608 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:26Z","lastTransitionTime":"2025-11-24T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.387132 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.387198 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.387214 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.387239 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.387253 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:26Z","lastTransitionTime":"2025-11-24T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:26 crc kubenswrapper[4777]: E1124 17:02:26.410684 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:26Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.417587 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.417640 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.417659 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.417685 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.417702 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:26Z","lastTransitionTime":"2025-11-24T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:26 crc kubenswrapper[4777]: E1124 17:02:26.439047 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:26Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.445540 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.445608 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.445629 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.445658 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.445677 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:26Z","lastTransitionTime":"2025-11-24T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:26 crc kubenswrapper[4777]: E1124 17:02:26.468693 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:26Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.474125 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.474173 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.474187 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.474207 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.474221 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:26Z","lastTransitionTime":"2025-11-24T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:26 crc kubenswrapper[4777]: E1124 17:02:26.489071 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:26Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.494792 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.494851 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.494863 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.494886 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.494898 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:26Z","lastTransitionTime":"2025-11-24T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:26 crc kubenswrapper[4777]: E1124 17:02:26.508925 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:26Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:26 crc kubenswrapper[4777]: E1124 17:02:26.509122 4777 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.510907 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.510994 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.511013 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.511052 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.511065 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:26Z","lastTransitionTime":"2025-11-24T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.614420 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.614475 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.614493 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.614519 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.614542 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:26Z","lastTransitionTime":"2025-11-24T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.718335 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.718403 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.718420 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.718445 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.718462 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:26Z","lastTransitionTime":"2025-11-24T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.821640 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.821716 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.821741 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.821772 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.821799 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:26Z","lastTransitionTime":"2025-11-24T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.835760 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs\") pod \"network-metrics-daemon-kxqzn\" (UID: \"960e1bed-abef-4f45-9475-37eee6bd780a\") " pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:26 crc kubenswrapper[4777]: E1124 17:02:26.836056 4777 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:02:26 crc kubenswrapper[4777]: E1124 17:02:26.836148 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs podName:960e1bed-abef-4f45-9475-37eee6bd780a nodeName:}" failed. No retries permitted until 2025-11-24 17:03:30.836124202 +0000 UTC m=+168.995319291 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs") pod "network-metrics-daemon-kxqzn" (UID: "960e1bed-abef-4f45-9475-37eee6bd780a") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.924576 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.924652 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.924679 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.924712 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:26 crc kubenswrapper[4777]: I1124 17:02:26.924783 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:26Z","lastTransitionTime":"2025-11-24T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.027255 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.027293 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.027301 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.027318 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.027330 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:27Z","lastTransitionTime":"2025-11-24T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.131015 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.131089 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.131107 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.131136 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.131161 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:27Z","lastTransitionTime":"2025-11-24T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.235609 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.235705 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.235735 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.235770 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.235820 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:27Z","lastTransitionTime":"2025-11-24T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.244752 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:27 crc kubenswrapper[4777]: E1124 17:02:27.244937 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.340217 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.340271 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.340297 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.340331 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.340355 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:27Z","lastTransitionTime":"2025-11-24T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.443275 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.443338 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.443357 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.443387 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.443406 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:27Z","lastTransitionTime":"2025-11-24T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.546921 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.547003 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.547028 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.547057 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.547075 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:27Z","lastTransitionTime":"2025-11-24T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.652099 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.652205 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.652229 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.652263 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.652287 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:27Z","lastTransitionTime":"2025-11-24T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.755406 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.755475 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.755498 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.755525 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.755544 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:27Z","lastTransitionTime":"2025-11-24T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.859754 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.859825 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.859847 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.859875 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.859893 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:27Z","lastTransitionTime":"2025-11-24T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.969766 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.969843 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.969867 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.969900 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:27 crc kubenswrapper[4777]: I1124 17:02:27.969924 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:27Z","lastTransitionTime":"2025-11-24T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.072529 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.072574 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.072583 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.072601 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.072611 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:28Z","lastTransitionTime":"2025-11-24T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.175050 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.175105 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.175116 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.175140 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.175156 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:28Z","lastTransitionTime":"2025-11-24T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.244870 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.244911 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.245112 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:28 crc kubenswrapper[4777]: E1124 17:02:28.245336 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:28 crc kubenswrapper[4777]: E1124 17:02:28.245595 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:28 crc kubenswrapper[4777]: E1124 17:02:28.245679 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.262510 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.278871 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.278949 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.279040 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.279072 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.279093 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:28Z","lastTransitionTime":"2025-11-24T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.382200 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.382289 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.382307 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.382335 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.382355 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:28Z","lastTransitionTime":"2025-11-24T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.485630 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.485688 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.485704 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.485730 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.485751 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:28Z","lastTransitionTime":"2025-11-24T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.588925 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.589030 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.589049 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.589076 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.589096 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:28Z","lastTransitionTime":"2025-11-24T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.691786 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.691871 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.691895 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.691928 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.691951 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:28Z","lastTransitionTime":"2025-11-24T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.794949 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.795079 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.795106 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.795144 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.795173 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:28Z","lastTransitionTime":"2025-11-24T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.904347 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.904439 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.904502 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.904533 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:28 crc kubenswrapper[4777]: I1124 17:02:28.904590 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:28Z","lastTransitionTime":"2025-11-24T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.008387 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.008442 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.008456 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.008477 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.008492 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:29Z","lastTransitionTime":"2025-11-24T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.111028 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.111083 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.111100 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.111124 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.111142 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:29Z","lastTransitionTime":"2025-11-24T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.214574 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.214660 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.214684 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.214716 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.214740 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:29Z","lastTransitionTime":"2025-11-24T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.244859 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:29 crc kubenswrapper[4777]: E1124 17:02:29.245113 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.318285 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.318343 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.318359 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.318381 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.318398 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:29Z","lastTransitionTime":"2025-11-24T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.421177 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.421254 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.421278 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.421310 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.421332 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:29Z","lastTransitionTime":"2025-11-24T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.525043 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.525126 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.525150 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.525183 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.525206 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:29Z","lastTransitionTime":"2025-11-24T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.628747 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.628832 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.628857 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.628889 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.628908 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:29Z","lastTransitionTime":"2025-11-24T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.732775 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.732868 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.732891 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.732924 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.732946 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:29Z","lastTransitionTime":"2025-11-24T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.837141 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.837236 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.837263 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.837292 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.837311 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:29Z","lastTransitionTime":"2025-11-24T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.941667 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.941743 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.941763 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.941792 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:29 crc kubenswrapper[4777]: I1124 17:02:29.941814 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:29Z","lastTransitionTime":"2025-11-24T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.045606 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.045679 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.045697 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.045724 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.045743 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:30Z","lastTransitionTime":"2025-11-24T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.149256 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.149340 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.149359 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.149542 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.149571 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:30Z","lastTransitionTime":"2025-11-24T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.244245 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.244317 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.244370 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:30 crc kubenswrapper[4777]: E1124 17:02:30.244512 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:30 crc kubenswrapper[4777]: E1124 17:02:30.244645 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:30 crc kubenswrapper[4777]: E1124 17:02:30.244911 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.252948 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.253076 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.253100 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.253126 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.253148 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:30Z","lastTransitionTime":"2025-11-24T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.356358 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.356408 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.356419 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.356437 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.356449 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:30Z","lastTransitionTime":"2025-11-24T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.459234 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.459304 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.459326 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.459351 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.459370 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:30Z","lastTransitionTime":"2025-11-24T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.562495 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.562606 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.562625 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.562655 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.562672 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:30Z","lastTransitionTime":"2025-11-24T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.669066 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.669253 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.669282 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.669315 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.669339 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:30Z","lastTransitionTime":"2025-11-24T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.772572 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.772655 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.772676 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.772704 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.772725 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:30Z","lastTransitionTime":"2025-11-24T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.876093 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.876157 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.876175 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.876201 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.876220 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:30Z","lastTransitionTime":"2025-11-24T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.980379 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.980465 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.980497 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.980532 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:30 crc kubenswrapper[4777]: I1124 17:02:30.980560 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:30Z","lastTransitionTime":"2025-11-24T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.084502 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.084573 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.084592 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.084621 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.084640 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:31Z","lastTransitionTime":"2025-11-24T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.188550 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.188638 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.188662 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.188697 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.188722 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:31Z","lastTransitionTime":"2025-11-24T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.245317 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:31 crc kubenswrapper[4777]: E1124 17:02:31.245551 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.291332 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.291385 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.291400 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.291419 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.291432 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:31Z","lastTransitionTime":"2025-11-24T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.394747 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.394824 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.394843 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.394872 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.394893 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:31Z","lastTransitionTime":"2025-11-24T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.498720 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.498816 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.498841 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.498872 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.498898 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:31Z","lastTransitionTime":"2025-11-24T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.601818 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.601858 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.601867 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.601880 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.601891 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:31Z","lastTransitionTime":"2025-11-24T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.705755 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.705802 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.705818 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.705838 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.705850 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:31Z","lastTransitionTime":"2025-11-24T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.808818 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.808910 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.808933 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.809022 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.809049 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:31Z","lastTransitionTime":"2025-11-24T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.911837 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.911883 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.911893 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.911912 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:31 crc kubenswrapper[4777]: I1124 17:02:31.911925 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:31Z","lastTransitionTime":"2025-11-24T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.014653 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.014717 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.014739 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.014768 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.014796 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:32Z","lastTransitionTime":"2025-11-24T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.117927 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.118047 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.118074 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.118105 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.118127 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:32Z","lastTransitionTime":"2025-11-24T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.221432 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.221935 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.221959 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.222364 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.222690 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:32Z","lastTransitionTime":"2025-11-24T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.244644 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.244772 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.244814 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:32 crc kubenswrapper[4777]: E1124 17:02:32.245136 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:32 crc kubenswrapper[4777]: E1124 17:02:32.245229 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:32 crc kubenswrapper[4777]: E1124 17:02:32.244943 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.326294 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.326385 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.326409 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.326437 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.326457 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:32Z","lastTransitionTime":"2025-11-24T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.429864 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.429991 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.430019 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.430055 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.430084 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:32Z","lastTransitionTime":"2025-11-24T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.532995 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.533054 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.533070 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.533094 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.533115 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:32Z","lastTransitionTime":"2025-11-24T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.636769 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.636830 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.636846 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.636869 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.636888 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:32Z","lastTransitionTime":"2025-11-24T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.740381 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.740442 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.740455 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.740474 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.740488 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:32Z","lastTransitionTime":"2025-11-24T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.844057 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.844137 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.844174 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.844206 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.844229 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:32Z","lastTransitionTime":"2025-11-24T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.947896 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.947955 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.947995 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.948019 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:32 crc kubenswrapper[4777]: I1124 17:02:32.948037 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:32Z","lastTransitionTime":"2025-11-24T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.050625 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.050686 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.050708 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.050735 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.050754 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:33Z","lastTransitionTime":"2025-11-24T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.153845 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.153923 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.153942 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.154006 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.154025 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:33Z","lastTransitionTime":"2025-11-24T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.244414 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:33 crc kubenswrapper[4777]: E1124 17:02:33.244589 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.256676 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.256705 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.256714 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.256730 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.256740 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:33Z","lastTransitionTime":"2025-11-24T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.259716 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.276056 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.294214 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.310708 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d953828-f624-4e22-890b-8a3a704547a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b1c99062222e3f934e7f3ed4bf9a4deffde4ad69f12471b248c825e1adbafbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh8ww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vtgz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.341933 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9084c33d-bd72-470c-9f45-3b0b0f29aa19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:02:20Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-etcd/etcd\\\\\\\"}\\\\nI1124 17:02:20.267327 6959 services_controller.go:360] Finished syncing service etcd on namespace openshift-etcd for network=default : 1.404495ms\\\\nI1124 17:02:20.267342 6959 services_controller.go:356] Processing sync for service openshift-machine-api/cluster-autoscaler-operator for network=default\\\\nI1124 17:02:20.267160 6959 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1124 17:02:20.267364 6959 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 17:02:20.267426 6959 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:02:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kk9ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-djshf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.360065 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.360130 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.360477 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.360588 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.360765 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2d91d83-93ba-4997-8e57-bbefc542c8f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cf897c0ba74877e097920fcc8bedbe2b43f5992fbacce7dbf0b9b29cb5c3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9s6jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.360841 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:33Z","lastTransitionTime":"2025-11-24T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.378017 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bdc8331-01a9-4b35-9e2a-5d870e521fed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13ad8466a0f0b306dd124ed16530743baeb94e28b3e078e854859aaa093edf73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.393379 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db78007a-c827-48a8-8590-66fdcafb7717\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a551cc061d7105b0d4a9978872e1756073cc47290c5ed958f003ac770a2498f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034f342bd386d862217b259fae5bdf49b567ebbd9162b932f7d9b9f469a4e7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d0dc8c72b79ffb367d3a2120767ba87fa708f93aafa7950733826e935478ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd6568224ddbf83698ac6b0e23ebc789f64d4d2090c64d80a4681d9bb70668\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.411678 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.435467 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce07485-e819-45d1-a91d-4bf1623a4324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28de18308457395420df4e81cdf161d9d9ac556314406ea3262cc507e0f4a0d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20781d8ccbc51194ff38680acbe8181ce060fc27ed1f325f29ad4930ab2d862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9d4b3940b53c522f4aa724c358a07ecee46d1cad81c0fdf89341d907fa9d1c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623face43669f3219f0f637cca7846f2aba9b9e9d8cf7db3421cc2957565888e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c0a06bad16d6106369a8850964be8dd109d34575343b7ac0338e9d3d0d4b412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30151dcbd2b1d0fff1383db9a8ec0bb48d01a1f6e97c9a6a1401d8bfa4ea6db1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://862056a53541d918ed0a044146adb3fcdf347c4ac9de45695409ad670093d3f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:01:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7knbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtmnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.456023 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mfvtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5500f09-0e35-442c-87ae-8f280cd2edd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899132c850ac1d7af915f5cb557c347ff8caec6f210b43743788369f5852c7e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T17:01:58Z\\\",\\\"message\\\":\\\"2025-11-24T17:01:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7\\\\n2025-11-24T17:01:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_62c64f70-35fd-4ae9-80be-8ca3dfcc83b7 to /host/opt/cni/bin/\\\\n2025-11-24T17:01:13Z [verbose] multus-daemon started\\\\n2025-11-24T17:01:13Z [verbose] Readiness Indicator file check\\\\n2025-11-24T17:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mfvtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.465769 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.465800 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.465811 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.465825 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.465835 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:33Z","lastTransitionTime":"2025-11-24T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.471531 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zd25z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54ae5e11-7c57-4c4f-a35a-4654f24ecdc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168f29a24c6e3ab0289fce0f033f9efb65f5885bd8daa5bf351bcbf9a4d140e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llg6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zd25z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.490129 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bd214a-a09d-4bff-9cf3-5a6e1a4c863a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf19fb03f3d3efcef64176c5d8e7a998755e8aa656fb642269062d2f700f5f69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4dc9e6731279c3bb6cbe7511c42636860b9017057e068a3be82d5d9b4f1c160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wtnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6v6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.506526 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"960e1bed-abef-4f45-9475-37eee6bd780a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2vm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:01:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kxqzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.542154 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c00fd09d-3ed4-46c9-bcff-407342b1dd40\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbf4e41a8419e135180170ee0fccae2eeb52f5ddae2c2fc6363594ec31354fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356940adde9dbdf959d83687973a3e697156f8fe126c44a8597da63bfa75275d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd39c4a740bea1727c583d538f8008b020bf44860750851fed34535ee611cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d259e984ddb8f711944a5528f33f8c7e9a9f85e6cccc5e111804a85cf9ac642f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf8026bb6460f3b94c6e148f2f6e1cec97160b9b8ac26bd022f4a828faa5d38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://104e9b5dc0c02e1ea12feeb30c7d491adc0f671b881f68a0ce4e15304554b207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://104e9b5dc0c02e1ea12feeb30c7d491adc0f671b881f68a0ce4e15304554b207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700f6f49ed9507d3c3074fc19424595e0dbc2d95b1a51a76cfa65a0b8642e68e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://700f6f49ed9507d3c3074fc19424595e0dbc2d95b1a51a76cfa65a0b8642e68e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://55297085683d62c3b6c607487741e50eb2ac526c3304b140384bb33f85ce6d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55297085683d62c3b6c607487741e50eb2ac526c3304b140384bb33f85ce6d6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.563304 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3779d22750d9cb90f538768ef5d0f5956743291fa5064e5a9db505c544aab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.568451 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.568515 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.568534 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.568560 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.568578 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:33Z","lastTransitionTime":"2025-11-24T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.587457 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e14f4113bddfa8fb04b0760ee606d46ebfa36a400672b596c8c1a144a2525bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c5df66eb701dab52ed57469245e95f4c42d99139a0ed85336c012fd725fa4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.607308 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e20e8edb-433a-4d97-b010-324307ddf3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6b20acccae25414e52010f4c1f9306fcabf0e359c6d217c475ce2a3b913f6da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12288efd6814a5d7810e5b8ffe8b221e8778a7478566383adf7fc2e4368cf22d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://824cf781e6843cacb741e0a744211b5286d56b5f78f614e01929223b641a4211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7179e095c1f25b9f4076229717dfda304ee3b8c78c1cab76d084016b25f73fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.628621 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12a66737-aa97-4b83-a78f-ba048b5355c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a55ce575bdba72347f3efaf3a17238d443436582a5102aa4ff9280e926b5afdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8e025a774ef0eeb75307d19551b28c9eedbe5ed46e5d8d5a03b6bc1c1289453\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb0bb2a0b75d5611a6292c0eb289ac80600c12a2f7b307958b830af12feb1f6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7280474e7278415863e418f8d920b2a72d85a10db02f391d68216246d68196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673ba55c3d66eab66facf652ef5966206257e85b51f1a1a7d643cad5b215c8f2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"message\\\":\\\" (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 17:01:02.001597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 17:01:02.003197 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-6300844/tls.crt::/tmp/serving-cert-6300844/tls.key\\\\\\\"\\\\nI1124 17:01:07.239350 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 17:01:07.258235 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 17:01:07.258289 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 17:01:07.258340 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 17:01:07.258353 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 17:01:07.298010 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 17:01:07.298052 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298058 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 17:01:07.298063 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 17:01:07.298068 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1124 17:01:07.298044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 17:01:07.298072 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 17:01:07.298097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 17:01:07.300910 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d220071ba7aa4220529f7a02032ee4f3054969d0d32090b43f4c9b69d95b9633\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd2290eb8242f4b37eed5540681ace75346ccacd08cfb5f3ad0f5de7e0f941e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:33Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.679593 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.679642 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.679660 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.679688 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.679707 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:33Z","lastTransitionTime":"2025-11-24T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.782589 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.782655 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.782673 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.782700 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.782726 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:33Z","lastTransitionTime":"2025-11-24T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.884877 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.884923 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.884939 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.884960 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:33 crc kubenswrapper[4777]: I1124 17:02:33.885000 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:33Z","lastTransitionTime":"2025-11-24T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.004174 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.004255 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.004280 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.004312 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.004340 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:34Z","lastTransitionTime":"2025-11-24T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.107785 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.107836 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.107854 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.107878 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.107897 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:34Z","lastTransitionTime":"2025-11-24T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.211542 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.211623 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.211648 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.211680 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.211701 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:34Z","lastTransitionTime":"2025-11-24T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.245281 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.245391 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.245586 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:34 crc kubenswrapper[4777]: E1124 17:02:34.246157 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:34 crc kubenswrapper[4777]: E1124 17:02:34.246294 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:34 crc kubenswrapper[4777]: E1124 17:02:34.246470 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.246678 4777 scope.go:117] "RemoveContainer" containerID="1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e" Nov 24 17:02:34 crc kubenswrapper[4777]: E1124 17:02:34.246927 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\"" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.314473 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.314551 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.314579 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.314616 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.314645 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:34Z","lastTransitionTime":"2025-11-24T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.417598 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.417662 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.417686 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.417723 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.417747 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:34Z","lastTransitionTime":"2025-11-24T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.521647 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.521712 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.521734 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.521764 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.521784 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:34Z","lastTransitionTime":"2025-11-24T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.625102 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.625173 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.625190 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.625215 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.625233 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:34Z","lastTransitionTime":"2025-11-24T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.728743 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.728812 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.728829 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.728855 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.728879 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:34Z","lastTransitionTime":"2025-11-24T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.831924 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.832007 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.832025 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.832052 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.832073 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:34Z","lastTransitionTime":"2025-11-24T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.935905 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.936008 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.936026 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.936053 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:34 crc kubenswrapper[4777]: I1124 17:02:34.936071 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:34Z","lastTransitionTime":"2025-11-24T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.038916 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.038982 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.038996 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.039015 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.039029 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:35Z","lastTransitionTime":"2025-11-24T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.141690 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.141769 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.141802 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.141836 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.141859 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:35Z","lastTransitionTime":"2025-11-24T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.244615 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:35 crc kubenswrapper[4777]: E1124 17:02:35.244822 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.245297 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.245342 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.245359 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.245381 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.245400 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:35Z","lastTransitionTime":"2025-11-24T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.348711 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.348765 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.348781 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.348804 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.348822 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:35Z","lastTransitionTime":"2025-11-24T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.455683 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.455755 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.455772 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.455796 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.455816 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:35Z","lastTransitionTime":"2025-11-24T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.559305 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.559363 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.559380 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.559404 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.559424 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:35Z","lastTransitionTime":"2025-11-24T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.663719 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.663782 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.663802 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.663830 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.663849 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:35Z","lastTransitionTime":"2025-11-24T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.766636 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.766702 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.766725 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.766756 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.766782 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:35Z","lastTransitionTime":"2025-11-24T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.870572 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.870668 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.870692 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.870723 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.870747 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:35Z","lastTransitionTime":"2025-11-24T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.974133 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.974205 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.974229 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.974261 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:35 crc kubenswrapper[4777]: I1124 17:02:35.974284 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:35Z","lastTransitionTime":"2025-11-24T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.076548 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.076811 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.076844 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.076878 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.076901 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:36Z","lastTransitionTime":"2025-11-24T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.179664 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.179728 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.179748 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.179771 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.179790 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:36Z","lastTransitionTime":"2025-11-24T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.244313 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.244383 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.244321 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:36 crc kubenswrapper[4777]: E1124 17:02:36.244584 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:36 crc kubenswrapper[4777]: E1124 17:02:36.244682 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:36 crc kubenswrapper[4777]: E1124 17:02:36.245090 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.283267 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.283318 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.283335 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.283357 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.283374 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:36Z","lastTransitionTime":"2025-11-24T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.386652 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.386699 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.386715 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.386739 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.386757 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:36Z","lastTransitionTime":"2025-11-24T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.489408 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.489484 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.489524 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.489558 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.489583 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:36Z","lastTransitionTime":"2025-11-24T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.593156 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.593233 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.593252 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.593280 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.593299 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:36Z","lastTransitionTime":"2025-11-24T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.650080 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.650157 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.650190 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.650221 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.650304 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:36Z","lastTransitionTime":"2025-11-24T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:36 crc kubenswrapper[4777]: E1124 17:02:36.671218 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:36Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.676644 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.676695 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.676712 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.676737 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.676756 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:36Z","lastTransitionTime":"2025-11-24T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:36 crc kubenswrapper[4777]: E1124 17:02:36.697298 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:36Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.703634 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.703724 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.703746 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.703776 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.703798 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:36Z","lastTransitionTime":"2025-11-24T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:36 crc kubenswrapper[4777]: E1124 17:02:36.726823 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:36Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.744840 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.744905 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.744950 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.745009 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.745028 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:36Z","lastTransitionTime":"2025-11-24T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:36 crc kubenswrapper[4777]: E1124 17:02:36.773140 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:36Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.778921 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.779031 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.779053 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.779092 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.779113 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:36Z","lastTransitionTime":"2025-11-24T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:36 crc kubenswrapper[4777]: E1124 17:02:36.801425 4777 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T17:02:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"93f6ada4-92ef-4dea-97e3-58231cc70ccf\\\",\\\"systemUUID\\\":\\\"ab3e815f-373b-46b5-8331-bc00cff3390c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:36Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:36 crc kubenswrapper[4777]: E1124 17:02:36.801590 4777 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.803720 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.803765 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.803780 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.803802 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.803820 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:36Z","lastTransitionTime":"2025-11-24T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.907323 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.907392 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.907412 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.907437 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:36 crc kubenswrapper[4777]: I1124 17:02:36.907481 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:36Z","lastTransitionTime":"2025-11-24T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.010624 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.010678 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.010690 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.010708 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.010722 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:37Z","lastTransitionTime":"2025-11-24T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.113784 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.113857 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.113876 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.113905 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.113924 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:37Z","lastTransitionTime":"2025-11-24T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.217019 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.217102 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.217121 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.217144 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.217173 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:37Z","lastTransitionTime":"2025-11-24T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.244904 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:37 crc kubenswrapper[4777]: E1124 17:02:37.245147 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.320098 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.320704 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.321234 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.321462 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.321627 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:37Z","lastTransitionTime":"2025-11-24T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.425519 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.425620 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.425647 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.425685 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.425709 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:37Z","lastTransitionTime":"2025-11-24T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.529049 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.529336 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.529517 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.529684 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.529896 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:37Z","lastTransitionTime":"2025-11-24T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.634207 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.634270 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.634293 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.634324 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.634348 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:37Z","lastTransitionTime":"2025-11-24T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.737440 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.737490 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.737507 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.737532 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.737549 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:37Z","lastTransitionTime":"2025-11-24T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.839725 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.839769 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.839782 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.839801 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.839813 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:37Z","lastTransitionTime":"2025-11-24T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.942277 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.942356 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.942381 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.942413 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:37 crc kubenswrapper[4777]: I1124 17:02:37.942440 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:37Z","lastTransitionTime":"2025-11-24T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.045660 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.045724 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.045747 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.045769 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.045785 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:38Z","lastTransitionTime":"2025-11-24T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.149411 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.149490 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.149510 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.149540 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.149561 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:38Z","lastTransitionTime":"2025-11-24T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.244516 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.244624 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.244737 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:38 crc kubenswrapper[4777]: E1124 17:02:38.244942 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:38 crc kubenswrapper[4777]: E1124 17:02:38.245175 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:38 crc kubenswrapper[4777]: E1124 17:02:38.245413 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.253097 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.253165 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.253184 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.253212 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.253234 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:38Z","lastTransitionTime":"2025-11-24T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.357205 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.357271 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.357282 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.357305 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.357320 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:38Z","lastTransitionTime":"2025-11-24T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.461000 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.461057 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.461074 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.461102 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.461121 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:38Z","lastTransitionTime":"2025-11-24T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.564321 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.564390 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.564408 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.564436 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.564454 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:38Z","lastTransitionTime":"2025-11-24T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.685547 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.685612 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.685628 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.685653 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.685671 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:38Z","lastTransitionTime":"2025-11-24T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.789044 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.789119 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.789141 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.789168 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.789185 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:38Z","lastTransitionTime":"2025-11-24T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.892309 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.892367 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.892382 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.892404 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.892421 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:38Z","lastTransitionTime":"2025-11-24T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.995519 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.995604 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.995633 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.995668 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:38 crc kubenswrapper[4777]: I1124 17:02:38.995694 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:38Z","lastTransitionTime":"2025-11-24T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.098886 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.098946 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.098959 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.099006 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.099028 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:39Z","lastTransitionTime":"2025-11-24T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.203144 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.203195 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.203208 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.203226 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.203240 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:39Z","lastTransitionTime":"2025-11-24T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.244594 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:39 crc kubenswrapper[4777]: E1124 17:02:39.244795 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.305891 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.305931 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.305939 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.305952 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.305961 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:39Z","lastTransitionTime":"2025-11-24T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.408852 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.408913 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.408925 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.408942 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.408953 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:39Z","lastTransitionTime":"2025-11-24T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.512040 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.512089 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.512101 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.512118 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.512133 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:39Z","lastTransitionTime":"2025-11-24T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.614608 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.614650 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.614659 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.614672 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.614682 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:39Z","lastTransitionTime":"2025-11-24T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.717759 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.717799 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.717809 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.717822 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.717834 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:39Z","lastTransitionTime":"2025-11-24T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.820867 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.820952 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.821005 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.821038 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.821062 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:39Z","lastTransitionTime":"2025-11-24T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.922935 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.923036 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.923054 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.923081 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:39 crc kubenswrapper[4777]: I1124 17:02:39.923099 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:39Z","lastTransitionTime":"2025-11-24T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.026232 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.026332 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.026344 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.026359 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.026368 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:40Z","lastTransitionTime":"2025-11-24T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.129538 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.129609 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.129626 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.129651 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.129671 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:40Z","lastTransitionTime":"2025-11-24T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.234371 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.234428 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.234441 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.234460 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.234476 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:40Z","lastTransitionTime":"2025-11-24T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.244929 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.244936 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.245032 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:40 crc kubenswrapper[4777]: E1124 17:02:40.245124 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:40 crc kubenswrapper[4777]: E1124 17:02:40.245355 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:40 crc kubenswrapper[4777]: E1124 17:02:40.245472 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.337676 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.337752 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.337779 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.337817 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.337844 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:40Z","lastTransitionTime":"2025-11-24T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.441175 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.441249 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.441268 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.441297 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.441315 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:40Z","lastTransitionTime":"2025-11-24T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.544074 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.544145 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.544164 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.544192 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.544211 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:40Z","lastTransitionTime":"2025-11-24T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.646768 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.646837 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.646856 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.646882 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.646901 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:40Z","lastTransitionTime":"2025-11-24T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.749963 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.750054 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.750070 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.750095 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.750111 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:40Z","lastTransitionTime":"2025-11-24T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.853108 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.853163 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.853179 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.853203 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.853218 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:40Z","lastTransitionTime":"2025-11-24T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.956800 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.956868 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.956885 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.956910 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:40 crc kubenswrapper[4777]: I1124 17:02:40.956928 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:40Z","lastTransitionTime":"2025-11-24T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.059831 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.059893 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.059912 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.059936 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.059955 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:41Z","lastTransitionTime":"2025-11-24T17:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.163136 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.163196 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.163236 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.163269 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.163295 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:41Z","lastTransitionTime":"2025-11-24T17:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.245109 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:41 crc kubenswrapper[4777]: E1124 17:02:41.245321 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.266078 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.266129 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.266154 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.266184 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.266204 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:41Z","lastTransitionTime":"2025-11-24T17:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.369900 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.369995 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.370013 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.370040 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.370055 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:41Z","lastTransitionTime":"2025-11-24T17:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.472473 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.472542 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.472555 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.472580 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.472596 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:41Z","lastTransitionTime":"2025-11-24T17:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.576126 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.576190 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.576231 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.576267 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.576293 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:41Z","lastTransitionTime":"2025-11-24T17:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.678523 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.678580 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.678598 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.678621 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.678642 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:41Z","lastTransitionTime":"2025-11-24T17:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.782016 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.782113 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.782133 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.782158 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.782175 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:41Z","lastTransitionTime":"2025-11-24T17:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.885168 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.885230 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.885247 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.885272 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.885292 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:41Z","lastTransitionTime":"2025-11-24T17:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.987893 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.988025 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.988046 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.988071 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:41 crc kubenswrapper[4777]: I1124 17:02:41.988089 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:41Z","lastTransitionTime":"2025-11-24T17:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.090876 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.090928 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.090937 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.090949 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.090959 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:42Z","lastTransitionTime":"2025-11-24T17:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.193816 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.193902 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.193926 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.193956 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.194018 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:42Z","lastTransitionTime":"2025-11-24T17:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.244494 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.244584 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.244611 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:42 crc kubenswrapper[4777]: E1124 17:02:42.244647 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:42 crc kubenswrapper[4777]: E1124 17:02:42.244804 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:42 crc kubenswrapper[4777]: E1124 17:02:42.244889 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.296954 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.297712 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.297772 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.297812 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.297827 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:42Z","lastTransitionTime":"2025-11-24T17:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.400829 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.400889 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.400902 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.400921 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.400933 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:42Z","lastTransitionTime":"2025-11-24T17:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.504186 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.504245 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.504261 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.504286 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.504300 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:42Z","lastTransitionTime":"2025-11-24T17:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.607428 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.607514 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.607536 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.607565 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.607585 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:42Z","lastTransitionTime":"2025-11-24T17:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.710625 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.710685 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.710701 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.710755 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.710774 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:42Z","lastTransitionTime":"2025-11-24T17:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.813387 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.813470 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.813484 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.813505 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.813518 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:42Z","lastTransitionTime":"2025-11-24T17:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.916159 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.916228 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.916237 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.916286 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:42 crc kubenswrapper[4777]: I1124 17:02:42.916300 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:42Z","lastTransitionTime":"2025-11-24T17:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.019247 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.019305 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.019323 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.019346 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.019364 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:43Z","lastTransitionTime":"2025-11-24T17:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.122240 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.122287 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.122299 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.122321 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.122334 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:43Z","lastTransitionTime":"2025-11-24T17:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:43 crc kubenswrapper[4777]: E1124 17:02:43.222901 4777 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.246185 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:43 crc kubenswrapper[4777]: E1124 17:02:43.246409 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.264336 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.280306 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ad21f7186e61d52562658cc433ba9201747de45a23398cc047f0dd581ef780c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.300045 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.314842 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bdc8331-01a9-4b35-9e2a-5d870e521fed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13ad8466a0f0b306dd124ed16530743baeb94e28b3e078e854859aaa093edf73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://428c648979fb6b826dda25e00956f8888a1b2ce624f21c2f07993a7a209c976f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T17:00:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.332262 4777 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db78007a-c827-48a8-8590-66fdcafb7717\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a551cc061d7105b0d4a9978872e1756073cc47290c5ed958f003ac770a2498f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034f342bd386d862217b259fae5bdf49b567ebbd9162b932f7d9b9f469a4e7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d0dc8c72b79ffb367d3a2120767ba87fa708f93aafa7950733826e935478ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd6568224ddbf83698ac6b0e23ebc789f64d4d2090c64d80a4681d9bb70668\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T17:00:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T17:02:43Z is after 2025-08-24T17:21:41Z" Nov 24 17:02:43 crc kubenswrapper[4777]: E1124 17:02:43.426705 4777 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.437528 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podStartSLOduration=96.437511177 podStartE2EDuration="1m36.437511177s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:02:43.388294046 +0000 UTC m=+121.547489145" watchObservedRunningTime="2025-11-24 17:02:43.437511177 +0000 UTC m=+121.596706236" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.468366 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-h8twn" podStartSLOduration=96.46833813800001 podStartE2EDuration="1m36.468338138s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:02:43.450606278 +0000 UTC m=+121.609801327" watchObservedRunningTime="2025-11-24 17:02:43.468338138 +0000 UTC m=+121.627533187" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.499557 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=35.49952606 podStartE2EDuration="35.49952606s" podCreationTimestamp="2025-11-24 17:02:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:02:43.499307733 +0000 UTC m=+121.658502802" watchObservedRunningTime="2025-11-24 17:02:43.49952606 +0000 UTC m=+121.658721139" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.565299 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-xtmnl" podStartSLOduration=96.565250822 podStartE2EDuration="1m36.565250822s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:02:43.564530649 +0000 UTC m=+121.723725708" watchObservedRunningTime="2025-11-24 17:02:43.565250822 +0000 UTC m=+121.724445911" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.584817 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-mfvtr" podStartSLOduration=96.58480181 podStartE2EDuration="1m36.58480181s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:02:43.584145459 +0000 UTC m=+121.743340508" watchObservedRunningTime="2025-11-24 17:02:43.58480181 +0000 UTC m=+121.743996859" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.598439 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-zd25z" podStartSLOduration=96.598412678 podStartE2EDuration="1m36.598412678s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:02:43.597575811 +0000 UTC m=+121.756770880" watchObservedRunningTime="2025-11-24 17:02:43.598412678 +0000 UTC m=+121.757607767" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.630784 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6v6jn" podStartSLOduration=96.630761657 podStartE2EDuration="1m36.630761657s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:02:43.612205571 +0000 UTC m=+121.771400660" watchObservedRunningTime="2025-11-24 17:02:43.630761657 +0000 UTC m=+121.789956736" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.631255 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=68.631248473 podStartE2EDuration="1m8.631248473s" podCreationTimestamp="2025-11-24 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:02:43.630998065 +0000 UTC m=+121.790193124" watchObservedRunningTime="2025-11-24 17:02:43.631248473 +0000 UTC m=+121.790443562" Nov 24 17:02:43 crc kubenswrapper[4777]: I1124 17:02:43.647306 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=93.647292628 podStartE2EDuration="1m33.647292628s" podCreationTimestamp="2025-11-24 17:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:02:43.647152274 +0000 UTC m=+121.806347323" watchObservedRunningTime="2025-11-24 17:02:43.647292628 +0000 UTC m=+121.806487707" Nov 24 17:02:44 crc kubenswrapper[4777]: I1124 17:02:44.244951 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:44 crc kubenswrapper[4777]: I1124 17:02:44.244951 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:44 crc kubenswrapper[4777]: E1124 17:02:44.245192 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:44 crc kubenswrapper[4777]: I1124 17:02:44.244951 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:44 crc kubenswrapper[4777]: E1124 17:02:44.245334 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:44 crc kubenswrapper[4777]: E1124 17:02:44.245508 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:45 crc kubenswrapper[4777]: I1124 17:02:45.244901 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:45 crc kubenswrapper[4777]: E1124 17:02:45.245456 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:45 crc kubenswrapper[4777]: I1124 17:02:45.245724 4777 scope.go:117] "RemoveContainer" containerID="1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e" Nov 24 17:02:45 crc kubenswrapper[4777]: E1124 17:02:45.245959 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\"" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" Nov 24 17:02:46 crc kubenswrapper[4777]: I1124 17:02:46.056897 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mfvtr_a5500f09-0e35-442c-87ae-8f280cd2edd1/kube-multus/1.log" Nov 24 17:02:46 crc kubenswrapper[4777]: I1124 17:02:46.057893 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mfvtr_a5500f09-0e35-442c-87ae-8f280cd2edd1/kube-multus/0.log" Nov 24 17:02:46 crc kubenswrapper[4777]: I1124 17:02:46.057997 4777 generic.go:334] "Generic (PLEG): container finished" podID="a5500f09-0e35-442c-87ae-8f280cd2edd1" containerID="899132c850ac1d7af915f5cb557c347ff8caec6f210b43743788369f5852c7e8" exitCode=1 Nov 24 17:02:46 crc kubenswrapper[4777]: I1124 17:02:46.058056 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mfvtr" event={"ID":"a5500f09-0e35-442c-87ae-8f280cd2edd1","Type":"ContainerDied","Data":"899132c850ac1d7af915f5cb557c347ff8caec6f210b43743788369f5852c7e8"} Nov 24 17:02:46 crc kubenswrapper[4777]: I1124 17:02:46.058119 4777 scope.go:117] "RemoveContainer" containerID="dd8d349a753704d8d807a3b6dc01d5f75c4e9efe59ff1966d41b10343ee8c704" Nov 24 17:02:46 crc kubenswrapper[4777]: I1124 17:02:46.058804 4777 scope.go:117] "RemoveContainer" containerID="899132c850ac1d7af915f5cb557c347ff8caec6f210b43743788369f5852c7e8" Nov 24 17:02:46 crc kubenswrapper[4777]: E1124 17:02:46.059134 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-mfvtr_openshift-multus(a5500f09-0e35-442c-87ae-8f280cd2edd1)\"" pod="openshift-multus/multus-mfvtr" podUID="a5500f09-0e35-442c-87ae-8f280cd2edd1" Nov 24 17:02:46 crc kubenswrapper[4777]: I1124 17:02:46.136141 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=48.136114745 podStartE2EDuration="48.136114745s" podCreationTimestamp="2025-11-24 17:01:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:02:46.135592098 +0000 UTC m=+124.294787167" watchObservedRunningTime="2025-11-24 17:02:46.136114745 +0000 UTC m=+124.295309804" Nov 24 17:02:46 crc kubenswrapper[4777]: I1124 17:02:46.180102 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=18.180068128 podStartE2EDuration="18.180068128s" podCreationTimestamp="2025-11-24 17:02:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:02:46.15556789 +0000 UTC m=+124.314762949" watchObservedRunningTime="2025-11-24 17:02:46.180068128 +0000 UTC m=+124.339263217" Nov 24 17:02:46 crc kubenswrapper[4777]: I1124 17:02:46.244705 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:46 crc kubenswrapper[4777]: I1124 17:02:46.244789 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:46 crc kubenswrapper[4777]: I1124 17:02:46.244831 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:46 crc kubenswrapper[4777]: E1124 17:02:46.245017 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:46 crc kubenswrapper[4777]: E1124 17:02:46.245226 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:46 crc kubenswrapper[4777]: E1124 17:02:46.245389 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.064494 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mfvtr_a5500f09-0e35-442c-87ae-8f280cd2edd1/kube-multus/1.log" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.126609 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.126670 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.126688 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.126718 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.126741 4777 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T17:02:47Z","lastTransitionTime":"2025-11-24T17:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.180875 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5"] Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.181509 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.184636 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.184770 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.185126 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.186176 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.244370 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:47 crc kubenswrapper[4777]: E1124 17:02:47.244577 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.282253 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xdng5\" (UID: \"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.282291 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xdng5\" (UID: \"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.282371 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xdng5\" (UID: \"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.282442 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xdng5\" (UID: \"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.282473 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xdng5\" (UID: \"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.383748 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xdng5\" (UID: \"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.383818 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xdng5\" (UID: \"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.383954 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xdng5\" (UID: \"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.384062 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xdng5\" (UID: \"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.384090 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xdng5\" (UID: \"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.384106 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xdng5\" (UID: \"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.384777 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xdng5\" (UID: \"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.385690 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xdng5\" (UID: \"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.396774 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xdng5\" (UID: \"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.415956 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xdng5\" (UID: \"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:47 crc kubenswrapper[4777]: I1124 17:02:47.496659 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" Nov 24 17:02:48 crc kubenswrapper[4777]: I1124 17:02:48.071396 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" event={"ID":"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc","Type":"ContainerStarted","Data":"4f94452d089040907f0f1889ee80e9ab2a1648ed0f8ea4f88d4e9089f3598fc8"} Nov 24 17:02:48 crc kubenswrapper[4777]: I1124 17:02:48.072017 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" event={"ID":"1a3a6b6b-71b2-486a-9208-ae7df3a9b6dc","Type":"ContainerStarted","Data":"35a3cba6175d836bbe4ebe7e569e1cdfdc057538c0a33ee69915e236e23ff6a4"} Nov 24 17:02:48 crc kubenswrapper[4777]: I1124 17:02:48.245185 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:48 crc kubenswrapper[4777]: I1124 17:02:48.245231 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:48 crc kubenswrapper[4777]: E1124 17:02:48.245378 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:48 crc kubenswrapper[4777]: I1124 17:02:48.245649 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:48 crc kubenswrapper[4777]: E1124 17:02:48.245760 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:48 crc kubenswrapper[4777]: E1124 17:02:48.246151 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:48 crc kubenswrapper[4777]: E1124 17:02:48.428228 4777 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:02:49 crc kubenswrapper[4777]: I1124 17:02:49.245334 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:49 crc kubenswrapper[4777]: E1124 17:02:49.245568 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:50 crc kubenswrapper[4777]: I1124 17:02:50.244324 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:50 crc kubenswrapper[4777]: I1124 17:02:50.244389 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:50 crc kubenswrapper[4777]: I1124 17:02:50.244346 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:50 crc kubenswrapper[4777]: E1124 17:02:50.244480 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:50 crc kubenswrapper[4777]: E1124 17:02:50.244719 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:50 crc kubenswrapper[4777]: E1124 17:02:50.244915 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:51 crc kubenswrapper[4777]: I1124 17:02:51.244474 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:51 crc kubenswrapper[4777]: E1124 17:02:51.244850 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:52 crc kubenswrapper[4777]: I1124 17:02:52.244862 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:52 crc kubenswrapper[4777]: I1124 17:02:52.245004 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:52 crc kubenswrapper[4777]: E1124 17:02:52.245163 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:52 crc kubenswrapper[4777]: I1124 17:02:52.244889 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:52 crc kubenswrapper[4777]: E1124 17:02:52.245288 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:52 crc kubenswrapper[4777]: E1124 17:02:52.245436 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:53 crc kubenswrapper[4777]: I1124 17:02:53.245121 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:53 crc kubenswrapper[4777]: E1124 17:02:53.247124 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:53 crc kubenswrapper[4777]: E1124 17:02:53.428949 4777 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:02:54 crc kubenswrapper[4777]: I1124 17:02:54.245045 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:54 crc kubenswrapper[4777]: I1124 17:02:54.245146 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:54 crc kubenswrapper[4777]: E1124 17:02:54.245263 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:54 crc kubenswrapper[4777]: I1124 17:02:54.245363 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:54 crc kubenswrapper[4777]: E1124 17:02:54.245502 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:54 crc kubenswrapper[4777]: E1124 17:02:54.246143 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:55 crc kubenswrapper[4777]: I1124 17:02:55.244942 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:55 crc kubenswrapper[4777]: E1124 17:02:55.245189 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:56 crc kubenswrapper[4777]: I1124 17:02:56.245141 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:56 crc kubenswrapper[4777]: I1124 17:02:56.245226 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:56 crc kubenswrapper[4777]: I1124 17:02:56.245364 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:56 crc kubenswrapper[4777]: E1124 17:02:56.245546 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:56 crc kubenswrapper[4777]: E1124 17:02:56.246344 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:56 crc kubenswrapper[4777]: E1124 17:02:56.246457 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:56 crc kubenswrapper[4777]: I1124 17:02:56.247115 4777 scope.go:117] "RemoveContainer" containerID="1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e" Nov 24 17:02:56 crc kubenswrapper[4777]: E1124 17:02:56.247549 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-djshf_openshift-ovn-kubernetes(9084c33d-bd72-470c-9f45-3b0b0f29aa19)\"" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" Nov 24 17:02:57 crc kubenswrapper[4777]: I1124 17:02:57.244587 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:57 crc kubenswrapper[4777]: E1124 17:02:57.244776 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:02:58 crc kubenswrapper[4777]: I1124 17:02:58.245039 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:02:58 crc kubenswrapper[4777]: I1124 17:02:58.245078 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:02:58 crc kubenswrapper[4777]: I1124 17:02:58.245199 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:02:58 crc kubenswrapper[4777]: E1124 17:02:58.245318 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:02:58 crc kubenswrapper[4777]: E1124 17:02:58.245419 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:02:58 crc kubenswrapper[4777]: E1124 17:02:58.245582 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:02:58 crc kubenswrapper[4777]: E1124 17:02:58.430750 4777 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:02:59 crc kubenswrapper[4777]: I1124 17:02:59.244871 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:02:59 crc kubenswrapper[4777]: E1124 17:02:59.245275 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:03:00 crc kubenswrapper[4777]: I1124 17:03:00.244912 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:03:00 crc kubenswrapper[4777]: I1124 17:03:00.245093 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:03:00 crc kubenswrapper[4777]: E1124 17:03:00.245319 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:03:00 crc kubenswrapper[4777]: I1124 17:03:00.245373 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:03:00 crc kubenswrapper[4777]: E1124 17:03:00.245583 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:03:00 crc kubenswrapper[4777]: E1124 17:03:00.245767 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:03:00 crc kubenswrapper[4777]: I1124 17:03:00.246337 4777 scope.go:117] "RemoveContainer" containerID="899132c850ac1d7af915f5cb557c347ff8caec6f210b43743788369f5852c7e8" Nov 24 17:03:00 crc kubenswrapper[4777]: I1124 17:03:00.275816 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xdng5" podStartSLOduration=113.275792517 podStartE2EDuration="1m53.275792517s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:02:48.097902056 +0000 UTC m=+126.257097105" watchObservedRunningTime="2025-11-24 17:03:00.275792517 +0000 UTC m=+138.434987576" Nov 24 17:03:01 crc kubenswrapper[4777]: I1124 17:03:01.125627 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mfvtr_a5500f09-0e35-442c-87ae-8f280cd2edd1/kube-multus/1.log" Nov 24 17:03:01 crc kubenswrapper[4777]: I1124 17:03:01.125712 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mfvtr" event={"ID":"a5500f09-0e35-442c-87ae-8f280cd2edd1","Type":"ContainerStarted","Data":"f481bf73ae67258e05eff97b7b2653b611d0cd13cac8da9e0655f728f5bc954f"} Nov 24 17:03:01 crc kubenswrapper[4777]: I1124 17:03:01.244418 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:03:01 crc kubenswrapper[4777]: E1124 17:03:01.245004 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:03:02 crc kubenswrapper[4777]: I1124 17:03:02.245085 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:03:02 crc kubenswrapper[4777]: I1124 17:03:02.245155 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:03:02 crc kubenswrapper[4777]: E1124 17:03:02.245281 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:03:02 crc kubenswrapper[4777]: E1124 17:03:02.245384 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:03:02 crc kubenswrapper[4777]: I1124 17:03:02.245765 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:03:02 crc kubenswrapper[4777]: E1124 17:03:02.245892 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:03:03 crc kubenswrapper[4777]: I1124 17:03:03.244840 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:03:03 crc kubenswrapper[4777]: E1124 17:03:03.245745 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:03:03 crc kubenswrapper[4777]: E1124 17:03:03.431716 4777 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:03:04 crc kubenswrapper[4777]: I1124 17:03:04.244180 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:03:04 crc kubenswrapper[4777]: I1124 17:03:04.244211 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:03:04 crc kubenswrapper[4777]: I1124 17:03:04.244183 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:03:04 crc kubenswrapper[4777]: E1124 17:03:04.244330 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:03:04 crc kubenswrapper[4777]: E1124 17:03:04.244472 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:03:04 crc kubenswrapper[4777]: E1124 17:03:04.244689 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:03:05 crc kubenswrapper[4777]: I1124 17:03:05.244851 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:03:05 crc kubenswrapper[4777]: E1124 17:03:05.245069 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:03:06 crc kubenswrapper[4777]: I1124 17:03:06.244501 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:03:06 crc kubenswrapper[4777]: I1124 17:03:06.244598 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:03:06 crc kubenswrapper[4777]: E1124 17:03:06.244690 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:03:06 crc kubenswrapper[4777]: I1124 17:03:06.244620 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:03:06 crc kubenswrapper[4777]: E1124 17:03:06.244848 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:03:06 crc kubenswrapper[4777]: E1124 17:03:06.245388 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:03:07 crc kubenswrapper[4777]: I1124 17:03:07.244924 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:03:07 crc kubenswrapper[4777]: E1124 17:03:07.245128 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:03:08 crc kubenswrapper[4777]: I1124 17:03:08.244731 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:03:08 crc kubenswrapper[4777]: I1124 17:03:08.244738 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:03:08 crc kubenswrapper[4777]: I1124 17:03:08.244866 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:03:08 crc kubenswrapper[4777]: E1124 17:03:08.245053 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:03:08 crc kubenswrapper[4777]: E1124 17:03:08.245173 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:03:08 crc kubenswrapper[4777]: E1124 17:03:08.245308 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:03:08 crc kubenswrapper[4777]: E1124 17:03:08.432880 4777 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:03:09 crc kubenswrapper[4777]: I1124 17:03:09.245114 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:03:09 crc kubenswrapper[4777]: E1124 17:03:09.245309 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:03:10 crc kubenswrapper[4777]: I1124 17:03:10.244768 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:03:10 crc kubenswrapper[4777]: I1124 17:03:10.244858 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:03:10 crc kubenswrapper[4777]: I1124 17:03:10.244782 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:03:10 crc kubenswrapper[4777]: E1124 17:03:10.245343 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:03:10 crc kubenswrapper[4777]: I1124 17:03:10.245572 4777 scope.go:117] "RemoveContainer" containerID="1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e" Nov 24 17:03:10 crc kubenswrapper[4777]: E1124 17:03:10.245572 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:03:10 crc kubenswrapper[4777]: E1124 17:03:10.245805 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:03:11 crc kubenswrapper[4777]: I1124 17:03:11.163144 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/3.log" Nov 24 17:03:11 crc kubenswrapper[4777]: I1124 17:03:11.167733 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerStarted","Data":"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605"} Nov 24 17:03:11 crc kubenswrapper[4777]: I1124 17:03:11.168357 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:03:11 crc kubenswrapper[4777]: I1124 17:03:11.202920 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podStartSLOduration=124.202890016 podStartE2EDuration="2m4.202890016s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:11.20159924 +0000 UTC m=+149.360794329" watchObservedRunningTime="2025-11-24 17:03:11.202890016 +0000 UTC m=+149.362085075" Nov 24 17:03:11 crc kubenswrapper[4777]: I1124 17:03:11.244759 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:03:11 crc kubenswrapper[4777]: E1124 17:03:11.244951 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:03:11 crc kubenswrapper[4777]: I1124 17:03:11.281662 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kxqzn"] Nov 24 17:03:11 crc kubenswrapper[4777]: I1124 17:03:11.281861 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:03:11 crc kubenswrapper[4777]: E1124 17:03:11.282093 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:03:12 crc kubenswrapper[4777]: I1124 17:03:12.244482 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:03:12 crc kubenswrapper[4777]: I1124 17:03:12.244510 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:03:12 crc kubenswrapper[4777]: E1124 17:03:12.245038 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:03:12 crc kubenswrapper[4777]: E1124 17:03:12.245183 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:03:13 crc kubenswrapper[4777]: I1124 17:03:13.244888 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:03:13 crc kubenswrapper[4777]: I1124 17:03:13.245081 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:03:13 crc kubenswrapper[4777]: E1124 17:03:13.247122 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:03:13 crc kubenswrapper[4777]: E1124 17:03:13.247247 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:03:13 crc kubenswrapper[4777]: E1124 17:03:13.433588 4777 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:03:14 crc kubenswrapper[4777]: I1124 17:03:14.244720 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:03:14 crc kubenswrapper[4777]: E1124 17:03:14.244936 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:03:14 crc kubenswrapper[4777]: I1124 17:03:14.245351 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:03:14 crc kubenswrapper[4777]: E1124 17:03:14.245498 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:03:14 crc kubenswrapper[4777]: I1124 17:03:14.270663 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:03:14 crc kubenswrapper[4777]: I1124 17:03:14.270751 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:03:15 crc kubenswrapper[4777]: I1124 17:03:15.245050 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:03:15 crc kubenswrapper[4777]: I1124 17:03:15.245158 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:03:15 crc kubenswrapper[4777]: E1124 17:03:15.245232 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:03:15 crc kubenswrapper[4777]: E1124 17:03:15.245346 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:03:16 crc kubenswrapper[4777]: I1124 17:03:16.244286 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:03:16 crc kubenswrapper[4777]: I1124 17:03:16.244355 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:03:16 crc kubenswrapper[4777]: E1124 17:03:16.244507 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:03:16 crc kubenswrapper[4777]: E1124 17:03:16.244695 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:03:17 crc kubenswrapper[4777]: I1124 17:03:17.245034 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:03:17 crc kubenswrapper[4777]: I1124 17:03:17.245161 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:03:17 crc kubenswrapper[4777]: E1124 17:03:17.246277 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 17:03:17 crc kubenswrapper[4777]: E1124 17:03:17.246473 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kxqzn" podUID="960e1bed-abef-4f45-9475-37eee6bd780a" Nov 24 17:03:18 crc kubenswrapper[4777]: I1124 17:03:18.244670 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:03:18 crc kubenswrapper[4777]: I1124 17:03:18.244671 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:03:18 crc kubenswrapper[4777]: E1124 17:03:18.244911 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 17:03:18 crc kubenswrapper[4777]: E1124 17:03:18.245033 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 17:03:19 crc kubenswrapper[4777]: I1124 17:03:19.090721 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:19 crc kubenswrapper[4777]: I1124 17:03:19.090872 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:03:19 crc kubenswrapper[4777]: E1124 17:03:19.091013 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:05:21.090936717 +0000 UTC m=+279.250131796 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:19 crc kubenswrapper[4777]: E1124 17:03:19.091044 4777 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:03:19 crc kubenswrapper[4777]: E1124 17:03:19.091115 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:05:21.091093701 +0000 UTC m=+279.250288830 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 17:03:19 crc kubenswrapper[4777]: I1124 17:03:19.091142 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:03:19 crc kubenswrapper[4777]: I1124 17:03:19.091182 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:03:19 crc kubenswrapper[4777]: E1124 17:03:19.091314 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:03:19 crc kubenswrapper[4777]: E1124 17:03:19.091312 4777 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:03:19 crc kubenswrapper[4777]: E1124 17:03:19.091399 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 17:05:21.091377189 +0000 UTC m=+279.250572288 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 17:03:19 crc kubenswrapper[4777]: E1124 17:03:19.091334 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:03:19 crc kubenswrapper[4777]: E1124 17:03:19.091453 4777 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:03:19 crc kubenswrapper[4777]: E1124 17:03:19.091513 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 17:05:21.091499823 +0000 UTC m=+279.250694882 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:03:19 crc kubenswrapper[4777]: I1124 17:03:19.192547 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:03:19 crc kubenswrapper[4777]: E1124 17:03:19.192800 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 17:03:19 crc kubenswrapper[4777]: E1124 17:03:19.192836 4777 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 17:03:19 crc kubenswrapper[4777]: E1124 17:03:19.192856 4777 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:03:19 crc kubenswrapper[4777]: E1124 17:03:19.192930 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 17:05:21.192910896 +0000 UTC m=+279.352105975 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 17:03:19 crc kubenswrapper[4777]: I1124 17:03:19.245068 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:03:19 crc kubenswrapper[4777]: I1124 17:03:19.245112 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:03:19 crc kubenswrapper[4777]: I1124 17:03:19.247308 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 24 17:03:19 crc kubenswrapper[4777]: I1124 17:03:19.247573 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 24 17:03:19 crc kubenswrapper[4777]: I1124 17:03:19.248311 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 24 17:03:19 crc kubenswrapper[4777]: I1124 17:03:19.248388 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 24 17:03:20 crc kubenswrapper[4777]: I1124 17:03:20.244712 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:03:20 crc kubenswrapper[4777]: I1124 17:03:20.244742 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:03:20 crc kubenswrapper[4777]: I1124 17:03:20.248378 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 24 17:03:20 crc kubenswrapper[4777]: I1124 17:03:20.248493 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.871586 4777 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.915854 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-xtvn9"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.916459 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.920153 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.920608 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.920925 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qz2kt"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.921474 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.922067 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.922443 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.925769 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.932582 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.936006 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.932646 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.934128 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.958095 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.958263 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.958493 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.958598 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.958670 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.958731 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.958816 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.959029 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.959232 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.959615 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.959770 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.959889 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.960053 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.960181 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.960200 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.960058 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.960405 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.960523 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.960568 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-vk8pz"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.960648 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.960794 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.960998 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.961108 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.961196 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vk8pz" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.961238 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7hwq6"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.961694 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.962739 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-gcjhs"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.963054 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.963155 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.963760 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.964258 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.964484 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.965206 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.966849 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6b6gc"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.967715 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.968418 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.968875 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6b6gc" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.969159 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-jxbj9"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.976536 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.977609 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cw68j"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.977804 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.978895 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.979424 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qp69m"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.979786 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.980074 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.980172 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.980798 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.980830 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.980853 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.980990 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.981219 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.982065 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.984458 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.984789 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.985022 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.985174 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.985379 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.985502 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.986929 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.987230 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.987645 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.987864 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.991336 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.991381 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.991484 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.991574 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.991879 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.991916 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.991944 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.992134 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.992332 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.992333 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.993461 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.994162 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.994605 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-48b9w"] Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.995290 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-48b9w" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.997804 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.998080 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.998296 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.998518 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.998552 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.998771 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.998781 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.998905 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.999009 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.999051 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.999122 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.999013 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.999755 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.999810 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 24 17:03:27 crc kubenswrapper[4777]: I1124 17:03:27.999880 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:27.999997 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.046547 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.052956 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kd6cl"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.056703 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8062b545-bbdd-4a99-97c6-3682da1cb40f-serving-cert\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.056787 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8062b545-bbdd-4a99-97c6-3682da1cb40f-encryption-config\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.056821 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-etcd-client\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.056849 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92efc884-606f-4fca-9410-4856b39a985d-config\") pod \"machine-api-operator-5694c8668f-xtvn9\" (UID: \"92efc884-606f-4fca-9410-4856b39a985d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.056874 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8062b545-bbdd-4a99-97c6-3682da1cb40f-config\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.056916 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8062b545-bbdd-4a99-97c6-3682da1cb40f-audit-dir\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.056944 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8062b545-bbdd-4a99-97c6-3682da1cb40f-image-import-ca\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057010 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4w4r\" (UniqueName: \"kubernetes.io/projected/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-kube-api-access-b4w4r\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057081 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlqfr\" (UniqueName: \"kubernetes.io/projected/8062b545-bbdd-4a99-97c6-3682da1cb40f-kube-api-access-zlqfr\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057124 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q49hn\" (UniqueName: \"kubernetes.io/projected/92efc884-606f-4fca-9410-4856b39a985d-kube-api-access-q49hn\") pod \"machine-api-operator-5694c8668f-xtvn9\" (UID: \"92efc884-606f-4fca-9410-4856b39a985d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057177 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8062b545-bbdd-4a99-97c6-3682da1cb40f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057211 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8062b545-bbdd-4a99-97c6-3682da1cb40f-audit\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057233 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-audit-policies\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057266 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8062b545-bbdd-4a99-97c6-3682da1cb40f-etcd-client\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057443 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8062b545-bbdd-4a99-97c6-3682da1cb40f-etcd-serving-ca\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057480 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057576 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057578 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a08bdbf-d95e-4795-92be-f30739d00a6a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lhrpc\" (UID: \"5a08bdbf-d95e-4795-92be-f30739d00a6a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057625 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq2xk\" (UniqueName: \"kubernetes.io/projected/5a08bdbf-d95e-4795-92be-f30739d00a6a-kube-api-access-sq2xk\") pod \"openshift-controller-manager-operator-756b6f6bc6-lhrpc\" (UID: \"5a08bdbf-d95e-4795-92be-f30739d00a6a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057653 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8062b545-bbdd-4a99-97c6-3682da1cb40f-node-pullsecrets\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057683 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-audit-dir\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057709 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a08bdbf-d95e-4795-92be-f30739d00a6a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lhrpc\" (UID: \"5a08bdbf-d95e-4795-92be-f30739d00a6a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057791 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057816 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-encryption-config\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057859 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/92efc884-606f-4fca-9410-4856b39a985d-images\") pod \"machine-api-operator-5694c8668f-xtvn9\" (UID: \"92efc884-606f-4fca-9410-4856b39a985d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057888 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/92efc884-606f-4fca-9410-4856b39a985d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-xtvn9\" (UID: \"92efc884-606f-4fca-9410-4856b39a985d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.057910 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-serving-cert\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.058063 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.058291 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.058451 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.058595 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.058743 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.058885 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.059125 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.059427 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.059570 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.063117 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.063342 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.063432 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.064213 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.064280 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.064300 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.064464 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.064536 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.064590 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.064620 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.064734 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.064947 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.065072 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.065119 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.065079 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.065348 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.065501 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.065778 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.065866 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.066043 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.066540 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zmdsl"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.066634 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.066997 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zmdsl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.067874 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.070278 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.070364 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-chvs8"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.071218 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-chvs8" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.074153 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8fgbp"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.074780 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.074814 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9qdq5"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.075374 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-9qdq5" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.077259 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.079765 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.081789 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.090274 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.090328 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.090622 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.090618 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.090745 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.090906 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.091730 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-82dhj"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.091925 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.091999 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.094755 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-82dhj" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.094441 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.098056 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.100449 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.101988 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.102744 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.102837 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.103602 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.103773 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.105959 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.107106 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6tcwx"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.108206 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.108290 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.116459 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.118072 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.118363 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.118701 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wsl2k"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.119048 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.119225 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.123074 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.123491 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-9cm8z"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.123820 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-xtvn9"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.124067 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.124102 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qz2kt"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.124100 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.124319 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.125680 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.126704 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-wbx6n"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.127438 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-wbx6n" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.127806 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cw68j"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.128913 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-jxbj9"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.130919 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.131845 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.133011 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6b6gc"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.134847 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vk8pz"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.137900 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.139165 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.139343 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.140652 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-chvs8"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.142958 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zmdsl"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.148162 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7hwq6"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.149608 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qp69m"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.150941 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kd6cl"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.154415 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-gcjhs"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.155626 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9qdq5"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.156693 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.157718 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158351 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158456 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158493 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13c6ce51-e4c2-4397-8cd6-f101232d4bed-serving-cert\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158517 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2fd0dd98-6b46-4717-80c3-32d9502cae19-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-f6tp2\" (UID: \"2fd0dd98-6b46-4717-80c3-32d9502cae19\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158550 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-encryption-config\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158571 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/92efc884-606f-4fca-9410-4856b39a985d-images\") pod \"machine-api-operator-5694c8668f-xtvn9\" (UID: \"92efc884-606f-4fca-9410-4856b39a985d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158592 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/92efc884-606f-4fca-9410-4856b39a985d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-xtvn9\" (UID: \"92efc884-606f-4fca-9410-4856b39a985d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158614 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158635 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nzmv\" (UniqueName: \"kubernetes.io/projected/698429dc-65d4-4b5f-b773-544f52c276c7-kube-api-access-5nzmv\") pod \"packageserver-d55dfcdfc-rrgd4\" (UID: \"698429dc-65d4-4b5f-b773-544f52c276c7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158656 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158708 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-serving-cert\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158731 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhv4w\" (UniqueName: \"kubernetes.io/projected/13c6ce51-e4c2-4397-8cd6-f101232d4bed-kube-api-access-zhv4w\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158751 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-trusted-ca-bundle\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158782 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2148eaac-ce88-4048-9ae9-9ff6df2d4747-config\") pod \"route-controller-manager-6576b87f9c-zzd7t\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158805 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158827 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-client-ca\") pod \"controller-manager-879f6c89f-kd6cl\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158845 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-service-ca\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158867 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-752sw\" (UniqueName: \"kubernetes.io/projected/10516d2d-106f-4155-ab2c-c0a302ca5cbf-kube-api-access-752sw\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158890 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8062b545-bbdd-4a99-97c6-3682da1cb40f-serving-cert\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158913 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6r9w\" (UniqueName: \"kubernetes.io/projected/2148eaac-ce88-4048-9ae9-9ff6df2d4747-kube-api-access-j6r9w\") pod \"route-controller-manager-6576b87f9c-zzd7t\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158935 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/698429dc-65d4-4b5f-b773-544f52c276c7-apiservice-cert\") pod \"packageserver-d55dfcdfc-rrgd4\" (UID: \"698429dc-65d4-4b5f-b773-544f52c276c7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158957 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkdnr\" (UniqueName: \"kubernetes.io/projected/2fd0dd98-6b46-4717-80c3-32d9502cae19-kube-api-access-wkdnr\") pod \"machine-config-controller-84d6567774-f6tp2\" (UID: \"2fd0dd98-6b46-4717-80c3-32d9502cae19\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.158998 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-oauth-config\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159017 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/13c6ce51-e4c2-4397-8cd6-f101232d4bed-etcd-service-ca\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159037 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0f99ce69-4367-49eb-84c8-534959fa1735-metrics-tls\") pod \"ingress-operator-5b745b69d9-d6bzx\" (UID: \"0f99ce69-4367-49eb-84c8-534959fa1735\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159054 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2fd0dd98-6b46-4717-80c3-32d9502cae19-proxy-tls\") pod \"machine-config-controller-84d6567774-f6tp2\" (UID: \"2fd0dd98-6b46-4717-80c3-32d9502cae19\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159075 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/698429dc-65d4-4b5f-b773-544f52c276c7-webhook-cert\") pod \"packageserver-d55dfcdfc-rrgd4\" (UID: \"698429dc-65d4-4b5f-b773-544f52c276c7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159093 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ff9e4751-a4e2-43c7-b75f-29ea3f7beca9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-s86qg\" (UID: \"ff9e4751-a4e2-43c7-b75f-29ea3f7beca9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159110 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sqqw\" (UniqueName: \"kubernetes.io/projected/54a60a6f-27d1-44d8-8372-512191c72f89-kube-api-access-8sqqw\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159128 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9773b6b0-ffe7-40ed-b714-c7266b46e298-serving-cert\") pod \"controller-manager-879f6c89f-kd6cl\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159144 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8062b545-bbdd-4a99-97c6-3682da1cb40f-encryption-config\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159161 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-etcd-client\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159188 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159245 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktw2z\" (UniqueName: \"kubernetes.io/projected/0f99ce69-4367-49eb-84c8-534959fa1735-kube-api-access-ktw2z\") pod \"ingress-operator-5b745b69d9-d6bzx\" (UID: \"0f99ce69-4367-49eb-84c8-534959fa1735\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159267 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-serving-cert\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159311 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92efc884-606f-4fca-9410-4856b39a985d-config\") pod \"machine-api-operator-5694c8668f-xtvn9\" (UID: \"92efc884-606f-4fca-9410-4856b39a985d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159331 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8h6s\" (UniqueName: \"kubernetes.io/projected/ff9e4751-a4e2-43c7-b75f-29ea3f7beca9-kube-api-access-c8h6s\") pod \"cluster-image-registry-operator-dc59b4c8b-s86qg\" (UID: \"ff9e4751-a4e2-43c7-b75f-29ea3f7beca9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159353 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b8eb4c02-a9d2-416b-a363-da660f35d972-signing-key\") pod \"service-ca-9c57cc56f-48b9w\" (UID: \"b8eb4c02-a9d2-416b-a363-da660f35d972\") " pod="openshift-service-ca/service-ca-9c57cc56f-48b9w" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159390 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8062b545-bbdd-4a99-97c6-3682da1cb40f-config\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159410 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159429 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/eb260e29-56b9-4d87-9cee-7a8a3c336282-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cw68j\" (UID: \"eb260e29-56b9-4d87-9cee-7a8a3c336282\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159449 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159487 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8062b545-bbdd-4a99-97c6-3682da1cb40f-audit-dir\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159504 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/493ed584-ded6-4027-ab8d-3946ecc089a4-config\") pod \"kube-apiserver-operator-766d6c64bb-6dsxq\" (UID: \"493ed584-ded6-4027-ab8d-3946ecc089a4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159522 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/698429dc-65d4-4b5f-b773-544f52c276c7-tmpfs\") pod \"packageserver-d55dfcdfc-rrgd4\" (UID: \"698429dc-65d4-4b5f-b773-544f52c276c7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159562 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2148eaac-ce88-4048-9ae9-9ff6df2d4747-client-ca\") pod \"route-controller-manager-6576b87f9c-zzd7t\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159589 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54a60a6f-27d1-44d8-8372-512191c72f89-audit-dir\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159599 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/92efc884-606f-4fca-9410-4856b39a985d-images\") pod \"machine-api-operator-5694c8668f-xtvn9\" (UID: \"92efc884-606f-4fca-9410-4856b39a985d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159646 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8062b545-bbdd-4a99-97c6-3682da1cb40f-image-import-ca\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159854 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5fdd1057-81bd-43f9-90f7-a1c33bfafba9-auth-proxy-config\") pod \"machine-approver-56656f9798-4gxrm\" (UID: \"5fdd1057-81bd-43f9-90f7-a1c33bfafba9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159900 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b8eb4c02-a9d2-416b-a363-da660f35d972-signing-cabundle\") pod \"service-ca-9c57cc56f-48b9w\" (UID: \"b8eb4c02-a9d2-416b-a363-da660f35d972\") " pod="openshift-service-ca/service-ca-9c57cc56f-48b9w" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159919 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ca6224b-5e61-422a-b961-5ceeea6bef3a-trusted-ca\") pod \"console-operator-58897d9998-gcjhs\" (UID: \"2ca6224b-5e61-422a-b961-5ceeea6bef3a\") " pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159937 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/13c6ce51-e4c2-4397-8cd6-f101232d4bed-etcd-ca\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159979 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff9e4751-a4e2-43c7-b75f-29ea3f7beca9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-s86qg\" (UID: \"ff9e4751-a4e2-43c7-b75f-29ea3f7beca9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.159997 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8889\" (UniqueName: \"kubernetes.io/projected/0c79743f-3220-44ba-bc82-72ca21ee0daa-kube-api-access-m8889\") pod \"cluster-samples-operator-665b6dd947-6b6gc\" (UID: \"0c79743f-3220-44ba-bc82-72ca21ee0daa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6b6gc" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160027 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4w4r\" (UniqueName: \"kubernetes.io/projected/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-kube-api-access-b4w4r\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160288 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fdd1057-81bd-43f9-90f7-a1c33bfafba9-config\") pod \"machine-approver-56656f9798-4gxrm\" (UID: \"5fdd1057-81bd-43f9-90f7-a1c33bfafba9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160332 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0f99ce69-4367-49eb-84c8-534959fa1735-bound-sa-token\") pod \"ingress-operator-5b745b69d9-d6bzx\" (UID: \"0f99ce69-4367-49eb-84c8-534959fa1735\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160352 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-config\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160370 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160385 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/13c6ce51-e4c2-4397-8cd6-f101232d4bed-etcd-client\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160616 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0c79743f-3220-44ba-bc82-72ca21ee0daa-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6b6gc\" (UID: \"0c79743f-3220-44ba-bc82-72ca21ee0daa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6b6gc" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160638 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlqfr\" (UniqueName: \"kubernetes.io/projected/8062b545-bbdd-4a99-97c6-3682da1cb40f-kube-api-access-zlqfr\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160657 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttl95\" (UniqueName: \"kubernetes.io/projected/2ca6224b-5e61-422a-b961-5ceeea6bef3a-kube-api-access-ttl95\") pod \"console-operator-58897d9998-gcjhs\" (UID: \"2ca6224b-5e61-422a-b961-5ceeea6bef3a\") " pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160675 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-audit-policies\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160697 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q49hn\" (UniqueName: \"kubernetes.io/projected/92efc884-606f-4fca-9410-4856b39a985d-kube-api-access-q49hn\") pod \"machine-api-operator-5694c8668f-xtvn9\" (UID: \"92efc884-606f-4fca-9410-4856b39a985d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160734 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-config\") pod \"controller-manager-879f6c89f-kd6cl\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160751 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-oauth-serving-cert\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160804 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8062b545-bbdd-4a99-97c6-3682da1cb40f-image-import-ca\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160789 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpclm\" (UniqueName: \"kubernetes.io/projected/b8eb4c02-a9d2-416b-a363-da660f35d972-kube-api-access-zpclm\") pod \"service-ca-9c57cc56f-48b9w\" (UID: \"b8eb4c02-a9d2-416b-a363-da660f35d972\") " pod="openshift-service-ca/service-ca-9c57cc56f-48b9w" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160853 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160874 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77rt5\" (UniqueName: \"kubernetes.io/projected/9773b6b0-ffe7-40ed-b714-c7266b46e298-kube-api-access-77rt5\") pod \"controller-manager-879f6c89f-kd6cl\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160924 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8062b545-bbdd-4a99-97c6-3682da1cb40f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.160942 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2148eaac-ce88-4048-9ae9-9ff6df2d4747-serving-cert\") pod \"route-controller-manager-6576b87f9c-zzd7t\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161015 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb260e29-56b9-4d87-9cee-7a8a3c336282-serving-cert\") pod \"openshift-config-operator-7777fb866f-cw68j\" (UID: \"eb260e29-56b9-4d87-9cee-7a8a3c336282\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161032 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vrg5\" (UniqueName: \"kubernetes.io/projected/eb260e29-56b9-4d87-9cee-7a8a3c336282-kube-api-access-7vrg5\") pod \"openshift-config-operator-7777fb866f-cw68j\" (UID: \"eb260e29-56b9-4d87-9cee-7a8a3c336282\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161048 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161066 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-522qw\" (UniqueName: \"kubernetes.io/projected/5fdd1057-81bd-43f9-90f7-a1c33bfafba9-kube-api-access-522qw\") pod \"machine-approver-56656f9798-4gxrm\" (UID: \"5fdd1057-81bd-43f9-90f7-a1c33bfafba9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161081 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161097 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/493ed584-ded6-4027-ab8d-3946ecc089a4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6dsxq\" (UID: \"493ed584-ded6-4027-ab8d-3946ecc089a4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161113 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161153 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8062b545-bbdd-4a99-97c6-3682da1cb40f-audit\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161171 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/493ed584-ded6-4027-ab8d-3946ecc089a4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6dsxq\" (UID: \"493ed584-ded6-4027-ab8d-3946ecc089a4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161186 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-audit-policies\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161259 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cplxb\" (UniqueName: \"kubernetes.io/projected/90c90945-24a7-4f17-871c-7358792f0db5-kube-api-access-cplxb\") pod \"downloads-7954f5f757-vk8pz\" (UID: \"90c90945-24a7-4f17-871c-7358792f0db5\") " pod="openshift-console/downloads-7954f5f757-vk8pz" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161276 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ff9e4751-a4e2-43c7-b75f-29ea3f7beca9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-s86qg\" (UID: \"ff9e4751-a4e2-43c7-b75f-29ea3f7beca9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161316 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ca6224b-5e61-422a-b961-5ceeea6bef3a-config\") pod \"console-operator-58897d9998-gcjhs\" (UID: \"2ca6224b-5e61-422a-b961-5ceeea6bef3a\") " pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161332 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c6ce51-e4c2-4397-8cd6-f101232d4bed-config\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161354 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8062b545-bbdd-4a99-97c6-3682da1cb40f-etcd-serving-ca\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161742 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f99ce69-4367-49eb-84c8-534959fa1735-trusted-ca\") pod \"ingress-operator-5b745b69d9-d6bzx\" (UID: \"0f99ce69-4367-49eb-84c8-534959fa1735\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161779 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8062b545-bbdd-4a99-97c6-3682da1cb40f-etcd-client\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161796 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5fdd1057-81bd-43f9-90f7-a1c33bfafba9-machine-approver-tls\") pod \"machine-approver-56656f9798-4gxrm\" (UID: \"5fdd1057-81bd-43f9-90f7-a1c33bfafba9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161813 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161829 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a08bdbf-d95e-4795-92be-f30739d00a6a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lhrpc\" (UID: \"5a08bdbf-d95e-4795-92be-f30739d00a6a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.161893 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq2xk\" (UniqueName: \"kubernetes.io/projected/5a08bdbf-d95e-4795-92be-f30739d00a6a-kube-api-access-sq2xk\") pod \"openshift-controller-manager-operator-756b6f6bc6-lhrpc\" (UID: \"5a08bdbf-d95e-4795-92be-f30739d00a6a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.162036 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.162056 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8062b545-bbdd-4a99-97c6-3682da1cb40f-audit-dir\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.162067 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-82dhj"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.162161 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.162178 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8062b545-bbdd-4a99-97c6-3682da1cb40f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.162223 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.162871 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-audit-policies\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.163164 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92efc884-606f-4fca-9410-4856b39a985d-config\") pod \"machine-api-operator-5694c8668f-xtvn9\" (UID: \"92efc884-606f-4fca-9410-4856b39a985d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.163351 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8062b545-bbdd-4a99-97c6-3682da1cb40f-node-pullsecrets\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.163383 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8062b545-bbdd-4a99-97c6-3682da1cb40f-etcd-serving-ca\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.163447 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ca6224b-5e61-422a-b961-5ceeea6bef3a-serving-cert\") pod \"console-operator-58897d9998-gcjhs\" (UID: \"2ca6224b-5e61-422a-b961-5ceeea6bef3a\") " pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.163520 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a08bdbf-d95e-4795-92be-f30739d00a6a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lhrpc\" (UID: \"5a08bdbf-d95e-4795-92be-f30739d00a6a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.163720 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kd6cl\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.163767 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-audit-dir\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.163883 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-audit-dir\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.163902 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8062b545-bbdd-4a99-97c6-3682da1cb40f-config\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.163987 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8062b545-bbdd-4a99-97c6-3682da1cb40f-audit\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.164062 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8062b545-bbdd-4a99-97c6-3682da1cb40f-node-pullsecrets\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.164320 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.164469 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a08bdbf-d95e-4795-92be-f30739d00a6a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lhrpc\" (UID: \"5a08bdbf-d95e-4795-92be-f30739d00a6a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.165306 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8062b545-bbdd-4a99-97c6-3682da1cb40f-serving-cert\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.165395 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.165728 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-etcd-client\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.165849 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.166383 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/92efc884-606f-4fca-9410-4856b39a985d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-xtvn9\" (UID: \"92efc884-606f-4fca-9410-4856b39a985d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.166420 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8062b545-bbdd-4a99-97c6-3682da1cb40f-etcd-client\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.167088 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.167392 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-serving-cert\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.167799 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.168061 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8062b545-bbdd-4a99-97c6-3682da1cb40f-encryption-config\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.168767 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8fgbp"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.169706 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.170658 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-48b9w"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.171653 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.172599 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.173558 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.174352 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a08bdbf-d95e-4795-92be-f30739d00a6a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lhrpc\" (UID: \"5a08bdbf-d95e-4795-92be-f30739d00a6a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.174534 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-fsbnq"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.176074 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.175964 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-bcj2h"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.177155 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bcj2h" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.177414 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6tcwx"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.178237 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-encryption-config\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.178555 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.178780 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wsl2k"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.179818 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-fsbnq"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.180783 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.184220 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-wbx6n"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.184793 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-lzzxl"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.186073 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-lzzxl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.186155 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-lzzxl"] Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.198789 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.218908 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.259051 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.264556 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpclm\" (UniqueName: \"kubernetes.io/projected/b8eb4c02-a9d2-416b-a363-da660f35d972-kube-api-access-zpclm\") pod \"service-ca-9c57cc56f-48b9w\" (UID: \"b8eb4c02-a9d2-416b-a363-da660f35d972\") " pod="openshift-service-ca/service-ca-9c57cc56f-48b9w" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.264626 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-config\") pod \"controller-manager-879f6c89f-kd6cl\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.264657 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77rt5\" (UniqueName: \"kubernetes.io/projected/9773b6b0-ffe7-40ed-b714-c7266b46e298-kube-api-access-77rt5\") pod \"controller-manager-879f6c89f-kd6cl\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.264704 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2148eaac-ce88-4048-9ae9-9ff6df2d4747-serving-cert\") pod \"route-controller-manager-6576b87f9c-zzd7t\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.264722 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.264742 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.264776 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/493ed584-ded6-4027-ab8d-3946ecc089a4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6dsxq\" (UID: \"493ed584-ded6-4027-ab8d-3946ecc089a4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.264797 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af533a06-13ef-4813-979d-875ffa3639dc-service-ca-bundle\") pod \"authentication-operator-69f744f599-8fgbp\" (UID: \"af533a06-13ef-4813-979d-875ffa3639dc\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.264816 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ff9e4751-a4e2-43c7-b75f-29ea3f7beca9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-s86qg\" (UID: \"ff9e4751-a4e2-43c7-b75f-29ea3f7beca9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.264852 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfrbv\" (UniqueName: \"kubernetes.io/projected/903c1f59-fc19-48a1-941e-d67c125e0d42-kube-api-access-xfrbv\") pod \"service-ca-operator-777779d784-7dhx6\" (UID: \"903c1f59-fc19-48a1-941e-d67c125e0d42\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.264871 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f99ce69-4367-49eb-84c8-534959fa1735-trusted-ca\") pod \"ingress-operator-5b745b69d9-d6bzx\" (UID: \"0f99ce69-4367-49eb-84c8-534959fa1735\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.264917 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5fdd1057-81bd-43f9-90f7-a1c33bfafba9-machine-approver-tls\") pod \"machine-approver-56656f9798-4gxrm\" (UID: \"5fdd1057-81bd-43f9-90f7-a1c33bfafba9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.264939 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/194ce833-4172-4abf-929c-2a9c296a9675-config-volume\") pod \"dns-default-wbx6n\" (UID: \"194ce833-4172-4abf-929c-2a9c296a9675\") " pod="openshift-dns/dns-default-wbx6n" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.264960 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/903c1f59-fc19-48a1-941e-d67c125e0d42-serving-cert\") pod \"service-ca-operator-777779d784-7dhx6\" (UID: \"903c1f59-fc19-48a1-941e-d67c125e0d42\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265002 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3369a23a-c296-4fda-8318-85dd8cea9913-metrics-tls\") pod \"dns-operator-744455d44c-9qdq5\" (UID: \"3369a23a-c296-4fda-8318-85dd8cea9913\") " pod="openshift-dns-operator/dns-operator-744455d44c-9qdq5" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265043 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13c6ce51-e4c2-4397-8cd6-f101232d4bed-serving-cert\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265094 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af533a06-13ef-4813-979d-875ffa3639dc-config\") pod \"authentication-operator-69f744f599-8fgbp\" (UID: \"af533a06-13ef-4813-979d-875ffa3639dc\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265123 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86b3f820-9431-4cee-bcac-bb17c9a43fab-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9bj8l\" (UID: \"86b3f820-9431-4cee-bcac-bb17c9a43fab\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265160 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/73fe25f4-c8e5-4d1e-a26e-25a3f0689593-srv-cert\") pod \"olm-operator-6b444d44fb-dtp89\" (UID: \"73fe25f4-c8e5-4d1e-a26e-25a3f0689593\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265179 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhv4w\" (UniqueName: \"kubernetes.io/projected/13c6ce51-e4c2-4397-8cd6-f101232d4bed-kube-api-access-zhv4w\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265198 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-trusted-ca-bundle\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265240 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2148eaac-ce88-4048-9ae9-9ff6df2d4747-config\") pod \"route-controller-manager-6576b87f9c-zzd7t\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265256 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/73fe25f4-c8e5-4d1e-a26e-25a3f0689593-profile-collector-cert\") pod \"olm-operator-6b444d44fb-dtp89\" (UID: \"73fe25f4-c8e5-4d1e-a26e-25a3f0689593\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265273 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7ec3567f-1218-441b-b33d-054f0c6a2e2d-secret-volume\") pod \"collect-profiles-29400060-j98hw\" (UID: \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265369 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7510a625-89f5-43c1-9ea2-7b836c01c676-proxy-tls\") pod \"machine-config-operator-74547568cd-x2k8g\" (UID: \"7510a625-89f5-43c1-9ea2-7b836c01c676\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265403 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7510a625-89f5-43c1-9ea2-7b836c01c676-auth-proxy-config\") pod \"machine-config-operator-74547568cd-x2k8g\" (UID: \"7510a625-89f5-43c1-9ea2-7b836c01c676\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265420 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/404e867b-354a-45f7-b27a-10a7f3842555-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-chvs8\" (UID: \"404e867b-354a-45f7-b27a-10a7f3842555\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-chvs8" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265438 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn9ck\" (UniqueName: \"kubernetes.io/projected/7ec3567f-1218-441b-b33d-054f0c6a2e2d-kube-api-access-qn9ck\") pod \"collect-profiles-29400060-j98hw\" (UID: \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265470 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h689\" (UniqueName: \"kubernetes.io/projected/7510a625-89f5-43c1-9ea2-7b836c01c676-kube-api-access-4h689\") pod \"machine-config-operator-74547568cd-x2k8g\" (UID: \"7510a625-89f5-43c1-9ea2-7b836c01c676\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265499 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ff9e4751-a4e2-43c7-b75f-29ea3f7beca9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-s86qg\" (UID: \"ff9e4751-a4e2-43c7-b75f-29ea3f7beca9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265518 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8h6s\" (UniqueName: \"kubernetes.io/projected/ff9e4751-a4e2-43c7-b75f-29ea3f7beca9-kube-api-access-c8h6s\") pod \"cluster-image-registry-operator-dc59b4c8b-s86qg\" (UID: \"ff9e4751-a4e2-43c7-b75f-29ea3f7beca9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265552 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265572 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/eb260e29-56b9-4d87-9cee-7a8a3c336282-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cw68j\" (UID: \"eb260e29-56b9-4d87-9cee-7a8a3c336282\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265576 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265588 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265652 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/493ed584-ded6-4027-ab8d-3946ecc089a4-config\") pod \"kube-apiserver-operator-766d6c64bb-6dsxq\" (UID: \"493ed584-ded6-4027-ab8d-3946ecc089a4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265673 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2148eaac-ce88-4048-9ae9-9ff6df2d4747-client-ca\") pod \"route-controller-manager-6576b87f9c-zzd7t\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265689 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54a60a6f-27d1-44d8-8372-512191c72f89-audit-dir\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265708 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b8eb4c02-a9d2-416b-a363-da660f35d972-signing-cabundle\") pod \"service-ca-9c57cc56f-48b9w\" (UID: \"b8eb4c02-a9d2-416b-a363-da660f35d972\") " pod="openshift-service-ca/service-ca-9c57cc56f-48b9w" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265724 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/13c6ce51-e4c2-4397-8cd6-f101232d4bed-etcd-ca\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265743 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9892a320-3e31-48df-bf8d-98441dd429ca-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncqcl\" (UID: \"9892a320-3e31-48df-bf8d-98441dd429ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265772 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff9e4751-a4e2-43c7-b75f-29ea3f7beca9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-s86qg\" (UID: \"ff9e4751-a4e2-43c7-b75f-29ea3f7beca9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265790 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqqgh\" (UniqueName: \"kubernetes.io/projected/73fe25f4-c8e5-4d1e-a26e-25a3f0689593-kube-api-access-sqqgh\") pod \"olm-operator-6b444d44fb-dtp89\" (UID: \"73fe25f4-c8e5-4d1e-a26e-25a3f0689593\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265813 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0c79743f-3220-44ba-bc82-72ca21ee0daa-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6b6gc\" (UID: \"0c79743f-3220-44ba-bc82-72ca21ee0daa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6b6gc" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265829 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-config\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265831 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54a60a6f-27d1-44d8-8372-512191c72f89-audit-dir\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265848 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttl95\" (UniqueName: \"kubernetes.io/projected/2ca6224b-5e61-422a-b961-5ceeea6bef3a-kube-api-access-ttl95\") pod \"console-operator-58897d9998-gcjhs\" (UID: \"2ca6224b-5e61-422a-b961-5ceeea6bef3a\") " pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265901 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f140bda-5ace-46c7-a5a7-242b6bf58207-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5cdnv\" (UID: \"4f140bda-5ace-46c7-a5a7-242b6bf58207\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265939 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-audit-policies\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265960 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-oauth-serving-cert\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.265992 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/194ce833-4172-4abf-929c-2a9c296a9675-metrics-tls\") pod \"dns-default-wbx6n\" (UID: \"194ce833-4172-4abf-929c-2a9c296a9675\") " pod="openshift-dns/dns-default-wbx6n" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266010 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9892a320-3e31-48df-bf8d-98441dd429ca-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncqcl\" (UID: \"9892a320-3e31-48df-bf8d-98441dd429ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266028 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb260e29-56b9-4d87-9cee-7a8a3c336282-serving-cert\") pod \"openshift-config-operator-7777fb866f-cw68j\" (UID: \"eb260e29-56b9-4d87-9cee-7a8a3c336282\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266047 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vrg5\" (UniqueName: \"kubernetes.io/projected/eb260e29-56b9-4d87-9cee-7a8a3c336282-kube-api-access-7vrg5\") pod \"openshift-config-operator-7777fb866f-cw68j\" (UID: \"eb260e29-56b9-4d87-9cee-7a8a3c336282\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266063 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266082 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbvd5\" (UniqueName: \"kubernetes.io/projected/35059f02-39e8-40b0-8295-8230591cd996-kube-api-access-vbvd5\") pod \"control-plane-machine-set-operator-78cbb6b69f-zmdsl\" (UID: \"35059f02-39e8-40b0-8295-8230591cd996\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zmdsl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266103 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-522qw\" (UniqueName: \"kubernetes.io/projected/5fdd1057-81bd-43f9-90f7-a1c33bfafba9-kube-api-access-522qw\") pod \"machine-approver-56656f9798-4gxrm\" (UID: \"5fdd1057-81bd-43f9-90f7-a1c33bfafba9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266120 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/493ed584-ded6-4027-ab8d-3946ecc089a4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6dsxq\" (UID: \"493ed584-ded6-4027-ab8d-3946ecc089a4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266138 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266157 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86b3f820-9431-4cee-bcac-bb17c9a43fab-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9bj8l\" (UID: \"86b3f820-9431-4cee-bcac-bb17c9a43fab\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266176 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndw9c\" (UniqueName: \"kubernetes.io/projected/af533a06-13ef-4813-979d-875ffa3639dc-kube-api-access-ndw9c\") pod \"authentication-operator-69f744f599-8fgbp\" (UID: \"af533a06-13ef-4813-979d-875ffa3639dc\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266197 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cplxb\" (UniqueName: \"kubernetes.io/projected/90c90945-24a7-4f17-871c-7358792f0db5-kube-api-access-cplxb\") pod \"downloads-7954f5f757-vk8pz\" (UID: \"90c90945-24a7-4f17-871c-7358792f0db5\") " pod="openshift-console/downloads-7954f5f757-vk8pz" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266217 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ca6224b-5e61-422a-b961-5ceeea6bef3a-config\") pod \"console-operator-58897d9998-gcjhs\" (UID: \"2ca6224b-5e61-422a-b961-5ceeea6bef3a\") " pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266235 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c6ce51-e4c2-4397-8cd6-f101232d4bed-config\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266287 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ca6224b-5e61-422a-b961-5ceeea6bef3a-serving-cert\") pod \"console-operator-58897d9998-gcjhs\" (UID: \"2ca6224b-5e61-422a-b961-5ceeea6bef3a\") " pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266311 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af533a06-13ef-4813-979d-875ffa3639dc-serving-cert\") pod \"authentication-operator-69f744f599-8fgbp\" (UID: \"af533a06-13ef-4813-979d-875ffa3639dc\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266349 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kd6cl\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266367 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7510a625-89f5-43c1-9ea2-7b836c01c676-images\") pod \"machine-config-operator-74547568cd-x2k8g\" (UID: \"7510a625-89f5-43c1-9ea2-7b836c01c676\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266387 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266406 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2fd0dd98-6b46-4717-80c3-32d9502cae19-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-f6tp2\" (UID: \"2fd0dd98-6b46-4717-80c3-32d9502cae19\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266429 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nzmv\" (UniqueName: \"kubernetes.io/projected/698429dc-65d4-4b5f-b773-544f52c276c7-kube-api-access-5nzmv\") pod \"packageserver-d55dfcdfc-rrgd4\" (UID: \"698429dc-65d4-4b5f-b773-544f52c276c7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266446 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266463 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vcdh\" (UniqueName: \"kubernetes.io/projected/194ce833-4172-4abf-929c-2a9c296a9675-kube-api-access-7vcdh\") pod \"dns-default-wbx6n\" (UID: \"194ce833-4172-4abf-929c-2a9c296a9675\") " pod="openshift-dns/dns-default-wbx6n" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266481 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266501 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-client-ca\") pod \"controller-manager-879f6c89f-kd6cl\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266518 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-service-ca\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266537 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9892a320-3e31-48df-bf8d-98441dd429ca-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncqcl\" (UID: \"9892a320-3e31-48df-bf8d-98441dd429ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266555 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/35059f02-39e8-40b0-8295-8230591cd996-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zmdsl\" (UID: \"35059f02-39e8-40b0-8295-8230591cd996\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zmdsl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266575 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-752sw\" (UniqueName: \"kubernetes.io/projected/10516d2d-106f-4155-ab2c-c0a302ca5cbf-kube-api-access-752sw\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266592 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6r9w\" (UniqueName: \"kubernetes.io/projected/2148eaac-ce88-4048-9ae9-9ff6df2d4747-kube-api-access-j6r9w\") pod \"route-controller-manager-6576b87f9c-zzd7t\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266608 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vkxs\" (UniqueName: \"kubernetes.io/projected/86b3f820-9431-4cee-bcac-bb17c9a43fab-kube-api-access-5vkxs\") pod \"kube-storage-version-migrator-operator-b67b599dd-9bj8l\" (UID: \"86b3f820-9431-4cee-bcac-bb17c9a43fab\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266626 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/698429dc-65d4-4b5f-b773-544f52c276c7-apiservice-cert\") pod \"packageserver-d55dfcdfc-rrgd4\" (UID: \"698429dc-65d4-4b5f-b773-544f52c276c7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266642 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkdnr\" (UniqueName: \"kubernetes.io/projected/2fd0dd98-6b46-4717-80c3-32d9502cae19-kube-api-access-wkdnr\") pod \"machine-config-controller-84d6567774-f6tp2\" (UID: \"2fd0dd98-6b46-4717-80c3-32d9502cae19\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266658 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-oauth-config\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266666 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2148eaac-ce88-4048-9ae9-9ff6df2d4747-client-ca\") pod \"route-controller-manager-6576b87f9c-zzd7t\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266696 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8jv9\" (UniqueName: \"kubernetes.io/projected/3369a23a-c296-4fda-8318-85dd8cea9913-kube-api-access-n8jv9\") pod \"dns-operator-744455d44c-9qdq5\" (UID: \"3369a23a-c296-4fda-8318-85dd8cea9913\") " pod="openshift-dns-operator/dns-operator-744455d44c-9qdq5" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266713 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af533a06-13ef-4813-979d-875ffa3639dc-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8fgbp\" (UID: \"af533a06-13ef-4813-979d-875ffa3639dc\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266731 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/698429dc-65d4-4b5f-b773-544f52c276c7-webhook-cert\") pod \"packageserver-d55dfcdfc-rrgd4\" (UID: \"698429dc-65d4-4b5f-b773-544f52c276c7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266746 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/13c6ce51-e4c2-4397-8cd6-f101232d4bed-etcd-service-ca\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266762 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0f99ce69-4367-49eb-84c8-534959fa1735-metrics-tls\") pod \"ingress-operator-5b745b69d9-d6bzx\" (UID: \"0f99ce69-4367-49eb-84c8-534959fa1735\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266778 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2fd0dd98-6b46-4717-80c3-32d9502cae19-proxy-tls\") pod \"machine-config-controller-84d6567774-f6tp2\" (UID: \"2fd0dd98-6b46-4717-80c3-32d9502cae19\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266796 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hpzl\" (UniqueName: \"kubernetes.io/projected/4f140bda-5ace-46c7-a5a7-242b6bf58207-kube-api-access-2hpzl\") pod \"openshift-apiserver-operator-796bbdcf4f-5cdnv\" (UID: \"4f140bda-5ace-46c7-a5a7-242b6bf58207\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266814 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266841 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sqqw\" (UniqueName: \"kubernetes.io/projected/54a60a6f-27d1-44d8-8372-512191c72f89-kube-api-access-8sqqw\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266856 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9773b6b0-ffe7-40ed-b714-c7266b46e298-serving-cert\") pod \"controller-manager-879f6c89f-kd6cl\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266874 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ec3567f-1218-441b-b33d-054f0c6a2e2d-config-volume\") pod \"collect-profiles-29400060-j98hw\" (UID: \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266896 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b8eb4c02-a9d2-416b-a363-da660f35d972-signing-key\") pod \"service-ca-9c57cc56f-48b9w\" (UID: \"b8eb4c02-a9d2-416b-a363-da660f35d972\") " pod="openshift-service-ca/service-ca-9c57cc56f-48b9w" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266913 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktw2z\" (UniqueName: \"kubernetes.io/projected/0f99ce69-4367-49eb-84c8-534959fa1735-kube-api-access-ktw2z\") pod \"ingress-operator-5b745b69d9-d6bzx\" (UID: \"0f99ce69-4367-49eb-84c8-534959fa1735\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266931 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-serving-cert\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266950 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f140bda-5ace-46c7-a5a7-242b6bf58207-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5cdnv\" (UID: \"4f140bda-5ace-46c7-a5a7-242b6bf58207\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.266983 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ca6224b-5e61-422a-b961-5ceeea6bef3a-trusted-ca\") pod \"console-operator-58897d9998-gcjhs\" (UID: \"2ca6224b-5e61-422a-b961-5ceeea6bef3a\") " pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.267019 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5fdd1057-81bd-43f9-90f7-a1c33bfafba9-auth-proxy-config\") pod \"machine-approver-56656f9798-4gxrm\" (UID: \"5fdd1057-81bd-43f9-90f7-a1c33bfafba9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.267036 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/698429dc-65d4-4b5f-b773-544f52c276c7-tmpfs\") pod \"packageserver-d55dfcdfc-rrgd4\" (UID: \"698429dc-65d4-4b5f-b773-544f52c276c7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.267060 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8889\" (UniqueName: \"kubernetes.io/projected/0c79743f-3220-44ba-bc82-72ca21ee0daa-kube-api-access-m8889\") pod \"cluster-samples-operator-665b6dd947-6b6gc\" (UID: \"0c79743f-3220-44ba-bc82-72ca21ee0daa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6b6gc" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.267170 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fdd1057-81bd-43f9-90f7-a1c33bfafba9-config\") pod \"machine-approver-56656f9798-4gxrm\" (UID: \"5fdd1057-81bd-43f9-90f7-a1c33bfafba9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.267188 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff9e4751-a4e2-43c7-b75f-29ea3f7beca9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-s86qg\" (UID: \"ff9e4751-a4e2-43c7-b75f-29ea3f7beca9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.267218 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0f99ce69-4367-49eb-84c8-534959fa1735-bound-sa-token\") pod \"ingress-operator-5b745b69d9-d6bzx\" (UID: \"0f99ce69-4367-49eb-84c8-534959fa1735\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.267260 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.267304 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/13c6ce51-e4c2-4397-8cd6-f101232d4bed-etcd-client\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.267380 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-trusted-ca-bundle\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.267690 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-config\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.268045 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b8eb4c02-a9d2-416b-a363-da660f35d972-signing-cabundle\") pod \"service-ca-9c57cc56f-48b9w\" (UID: \"b8eb4c02-a9d2-416b-a363-da660f35d972\") " pod="openshift-service-ca/service-ca-9c57cc56f-48b9w" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.268215 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/903c1f59-fc19-48a1-941e-d67c125e0d42-config\") pod \"service-ca-operator-777779d784-7dhx6\" (UID: \"903c1f59-fc19-48a1-941e-d67c125e0d42\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.268292 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j4wk\" (UniqueName: \"kubernetes.io/projected/404e867b-354a-45f7-b27a-10a7f3842555-kube-api-access-7j4wk\") pod \"multus-admission-controller-857f4d67dd-chvs8\" (UID: \"404e867b-354a-45f7-b27a-10a7f3842555\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-chvs8" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.268377 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-audit-policies\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.268522 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/493ed584-ded6-4027-ab8d-3946ecc089a4-config\") pod \"kube-apiserver-operator-766d6c64bb-6dsxq\" (UID: \"493ed584-ded6-4027-ab8d-3946ecc089a4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.268544 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2148eaac-ce88-4048-9ae9-9ff6df2d4747-config\") pod \"route-controller-manager-6576b87f9c-zzd7t\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.268932 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-oauth-serving-cert\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.268944 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ca6224b-5e61-422a-b961-5ceeea6bef3a-config\") pod \"console-operator-58897d9998-gcjhs\" (UID: \"2ca6224b-5e61-422a-b961-5ceeea6bef3a\") " pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.269245 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-service-ca\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.269362 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fdd1057-81bd-43f9-90f7-a1c33bfafba9-config\") pod \"machine-approver-56656f9798-4gxrm\" (UID: \"5fdd1057-81bd-43f9-90f7-a1c33bfafba9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.269649 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f99ce69-4367-49eb-84c8-534959fa1735-trusted-ca\") pod \"ingress-operator-5b745b69d9-d6bzx\" (UID: \"0f99ce69-4367-49eb-84c8-534959fa1735\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.269934 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.270040 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13c6ce51-e4c2-4397-8cd6-f101232d4bed-serving-cert\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.270384 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/698429dc-65d4-4b5f-b773-544f52c276c7-tmpfs\") pod \"packageserver-d55dfcdfc-rrgd4\" (UID: \"698429dc-65d4-4b5f-b773-544f52c276c7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.270536 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/13c6ce51-e4c2-4397-8cd6-f101232d4bed-etcd-service-ca\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.271382 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.271689 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5fdd1057-81bd-43f9-90f7-a1c33bfafba9-machine-approver-tls\") pod \"machine-approver-56656f9798-4gxrm\" (UID: \"5fdd1057-81bd-43f9-90f7-a1c33bfafba9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.272087 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.272266 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c6ce51-e4c2-4397-8cd6-f101232d4bed-config\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.272577 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0c79743f-3220-44ba-bc82-72ca21ee0daa-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6b6gc\" (UID: \"0c79743f-3220-44ba-bc82-72ca21ee0daa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6b6gc" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.272685 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5fdd1057-81bd-43f9-90f7-a1c33bfafba9-auth-proxy-config\") pod \"machine-approver-56656f9798-4gxrm\" (UID: \"5fdd1057-81bd-43f9-90f7-a1c33bfafba9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.273015 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/eb260e29-56b9-4d87-9cee-7a8a3c336282-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cw68j\" (UID: \"eb260e29-56b9-4d87-9cee-7a8a3c336282\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.273011 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.273332 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2fd0dd98-6b46-4717-80c3-32d9502cae19-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-f6tp2\" (UID: \"2fd0dd98-6b46-4717-80c3-32d9502cae19\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.273726 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/13c6ce51-e4c2-4397-8cd6-f101232d4bed-etcd-ca\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.273886 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb260e29-56b9-4d87-9cee-7a8a3c336282-serving-cert\") pod \"openshift-config-operator-7777fb866f-cw68j\" (UID: \"eb260e29-56b9-4d87-9cee-7a8a3c336282\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.273988 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ca6224b-5e61-422a-b961-5ceeea6bef3a-trusted-ca\") pod \"console-operator-58897d9998-gcjhs\" (UID: \"2ca6224b-5e61-422a-b961-5ceeea6bef3a\") " pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.274032 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.274763 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.275652 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.275793 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b8eb4c02-a9d2-416b-a363-da660f35d972-signing-key\") pod \"service-ca-9c57cc56f-48b9w\" (UID: \"b8eb4c02-a9d2-416b-a363-da660f35d972\") " pod="openshift-service-ca/service-ca-9c57cc56f-48b9w" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.275863 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2fd0dd98-6b46-4717-80c3-32d9502cae19-proxy-tls\") pod \"machine-config-controller-84d6567774-f6tp2\" (UID: \"2fd0dd98-6b46-4717-80c3-32d9502cae19\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.276229 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.276259 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/13c6ce51-e4c2-4397-8cd6-f101232d4bed-etcd-client\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.276395 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/493ed584-ded6-4027-ab8d-3946ecc089a4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6dsxq\" (UID: \"493ed584-ded6-4027-ab8d-3946ecc089a4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.276909 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/698429dc-65d4-4b5f-b773-544f52c276c7-webhook-cert\") pod \"packageserver-d55dfcdfc-rrgd4\" (UID: \"698429dc-65d4-4b5f-b773-544f52c276c7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.277230 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ff9e4751-a4e2-43c7-b75f-29ea3f7beca9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-s86qg\" (UID: \"ff9e4751-a4e2-43c7-b75f-29ea3f7beca9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.277296 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-serving-cert\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.277726 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/698429dc-65d4-4b5f-b773-544f52c276c7-apiservice-cert\") pod \"packageserver-d55dfcdfc-rrgd4\" (UID: \"698429dc-65d4-4b5f-b773-544f52c276c7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.277891 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0f99ce69-4367-49eb-84c8-534959fa1735-metrics-tls\") pod \"ingress-operator-5b745b69d9-d6bzx\" (UID: \"0f99ce69-4367-49eb-84c8-534959fa1735\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.278197 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.278437 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ca6224b-5e61-422a-b961-5ceeea6bef3a-serving-cert\") pod \"console-operator-58897d9998-gcjhs\" (UID: \"2ca6224b-5e61-422a-b961-5ceeea6bef3a\") " pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.278721 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-oauth-config\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.278991 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.280284 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.284685 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2148eaac-ce88-4048-9ae9-9ff6df2d4747-serving-cert\") pod \"route-controller-manager-6576b87f9c-zzd7t\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.298650 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.306726 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9773b6b0-ffe7-40ed-b714-c7266b46e298-serving-cert\") pod \"controller-manager-879f6c89f-kd6cl\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.318712 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.326523 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-config\") pod \"controller-manager-879f6c89f-kd6cl\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.338808 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.344130 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-client-ca\") pod \"controller-manager-879f6c89f-kd6cl\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369147 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ec3567f-1218-441b-b33d-054f0c6a2e2d-config-volume\") pod \"collect-profiles-29400060-j98hw\" (UID: \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369194 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hpzl\" (UniqueName: \"kubernetes.io/projected/4f140bda-5ace-46c7-a5a7-242b6bf58207-kube-api-access-2hpzl\") pod \"openshift-apiserver-operator-796bbdcf4f-5cdnv\" (UID: \"4f140bda-5ace-46c7-a5a7-242b6bf58207\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369234 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f140bda-5ace-46c7-a5a7-242b6bf58207-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5cdnv\" (UID: \"4f140bda-5ace-46c7-a5a7-242b6bf58207\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369286 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/903c1f59-fc19-48a1-941e-d67c125e0d42-config\") pod \"service-ca-operator-777779d784-7dhx6\" (UID: \"903c1f59-fc19-48a1-941e-d67c125e0d42\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369315 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j4wk\" (UniqueName: \"kubernetes.io/projected/404e867b-354a-45f7-b27a-10a7f3842555-kube-api-access-7j4wk\") pod \"multus-admission-controller-857f4d67dd-chvs8\" (UID: \"404e867b-354a-45f7-b27a-10a7f3842555\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-chvs8" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369369 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af533a06-13ef-4813-979d-875ffa3639dc-service-ca-bundle\") pod \"authentication-operator-69f744f599-8fgbp\" (UID: \"af533a06-13ef-4813-979d-875ffa3639dc\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369393 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfrbv\" (UniqueName: \"kubernetes.io/projected/903c1f59-fc19-48a1-941e-d67c125e0d42-kube-api-access-xfrbv\") pod \"service-ca-operator-777779d784-7dhx6\" (UID: \"903c1f59-fc19-48a1-941e-d67c125e0d42\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369443 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/194ce833-4172-4abf-929c-2a9c296a9675-config-volume\") pod \"dns-default-wbx6n\" (UID: \"194ce833-4172-4abf-929c-2a9c296a9675\") " pod="openshift-dns/dns-default-wbx6n" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369534 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/903c1f59-fc19-48a1-941e-d67c125e0d42-serving-cert\") pod \"service-ca-operator-777779d784-7dhx6\" (UID: \"903c1f59-fc19-48a1-941e-d67c125e0d42\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369590 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3369a23a-c296-4fda-8318-85dd8cea9913-metrics-tls\") pod \"dns-operator-744455d44c-9qdq5\" (UID: \"3369a23a-c296-4fda-8318-85dd8cea9913\") " pod="openshift-dns-operator/dns-operator-744455d44c-9qdq5" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369620 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af533a06-13ef-4813-979d-875ffa3639dc-config\") pod \"authentication-operator-69f744f599-8fgbp\" (UID: \"af533a06-13ef-4813-979d-875ffa3639dc\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369644 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86b3f820-9431-4cee-bcac-bb17c9a43fab-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9bj8l\" (UID: \"86b3f820-9431-4cee-bcac-bb17c9a43fab\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369669 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/73fe25f4-c8e5-4d1e-a26e-25a3f0689593-srv-cert\") pod \"olm-operator-6b444d44fb-dtp89\" (UID: \"73fe25f4-c8e5-4d1e-a26e-25a3f0689593\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369711 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/73fe25f4-c8e5-4d1e-a26e-25a3f0689593-profile-collector-cert\") pod \"olm-operator-6b444d44fb-dtp89\" (UID: \"73fe25f4-c8e5-4d1e-a26e-25a3f0689593\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369732 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7ec3567f-1218-441b-b33d-054f0c6a2e2d-secret-volume\") pod \"collect-profiles-29400060-j98hw\" (UID: \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369755 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7510a625-89f5-43c1-9ea2-7b836c01c676-proxy-tls\") pod \"machine-config-operator-74547568cd-x2k8g\" (UID: \"7510a625-89f5-43c1-9ea2-7b836c01c676\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369783 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369797 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7510a625-89f5-43c1-9ea2-7b836c01c676-auth-proxy-config\") pod \"machine-config-operator-74547568cd-x2k8g\" (UID: \"7510a625-89f5-43c1-9ea2-7b836c01c676\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369840 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/404e867b-354a-45f7-b27a-10a7f3842555-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-chvs8\" (UID: \"404e867b-354a-45f7-b27a-10a7f3842555\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-chvs8" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369868 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn9ck\" (UniqueName: \"kubernetes.io/projected/7ec3567f-1218-441b-b33d-054f0c6a2e2d-kube-api-access-qn9ck\") pod \"collect-profiles-29400060-j98hw\" (UID: \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369890 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h689\" (UniqueName: \"kubernetes.io/projected/7510a625-89f5-43c1-9ea2-7b836c01c676-kube-api-access-4h689\") pod \"machine-config-operator-74547568cd-x2k8g\" (UID: \"7510a625-89f5-43c1-9ea2-7b836c01c676\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369939 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9892a320-3e31-48df-bf8d-98441dd429ca-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncqcl\" (UID: \"9892a320-3e31-48df-bf8d-98441dd429ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.369986 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqqgh\" (UniqueName: \"kubernetes.io/projected/73fe25f4-c8e5-4d1e-a26e-25a3f0689593-kube-api-access-sqqgh\") pod \"olm-operator-6b444d44fb-dtp89\" (UID: \"73fe25f4-c8e5-4d1e-a26e-25a3f0689593\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.370019 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f140bda-5ace-46c7-a5a7-242b6bf58207-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5cdnv\" (UID: \"4f140bda-5ace-46c7-a5a7-242b6bf58207\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.370056 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/194ce833-4172-4abf-929c-2a9c296a9675-metrics-tls\") pod \"dns-default-wbx6n\" (UID: \"194ce833-4172-4abf-929c-2a9c296a9675\") " pod="openshift-dns/dns-default-wbx6n" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.370080 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9892a320-3e31-48df-bf8d-98441dd429ca-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncqcl\" (UID: \"9892a320-3e31-48df-bf8d-98441dd429ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.370136 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbvd5\" (UniqueName: \"kubernetes.io/projected/35059f02-39e8-40b0-8295-8230591cd996-kube-api-access-vbvd5\") pod \"control-plane-machine-set-operator-78cbb6b69f-zmdsl\" (UID: \"35059f02-39e8-40b0-8295-8230591cd996\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zmdsl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.370171 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86b3f820-9431-4cee-bcac-bb17c9a43fab-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9bj8l\" (UID: \"86b3f820-9431-4cee-bcac-bb17c9a43fab\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.370193 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndw9c\" (UniqueName: \"kubernetes.io/projected/af533a06-13ef-4813-979d-875ffa3639dc-kube-api-access-ndw9c\") pod \"authentication-operator-69f744f599-8fgbp\" (UID: \"af533a06-13ef-4813-979d-875ffa3639dc\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.370228 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af533a06-13ef-4813-979d-875ffa3639dc-serving-cert\") pod \"authentication-operator-69f744f599-8fgbp\" (UID: \"af533a06-13ef-4813-979d-875ffa3639dc\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.370267 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7510a625-89f5-43c1-9ea2-7b836c01c676-images\") pod \"machine-config-operator-74547568cd-x2k8g\" (UID: \"7510a625-89f5-43c1-9ea2-7b836c01c676\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.370300 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vcdh\" (UniqueName: \"kubernetes.io/projected/194ce833-4172-4abf-929c-2a9c296a9675-kube-api-access-7vcdh\") pod \"dns-default-wbx6n\" (UID: \"194ce833-4172-4abf-929c-2a9c296a9675\") " pod="openshift-dns/dns-default-wbx6n" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.370324 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9892a320-3e31-48df-bf8d-98441dd429ca-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncqcl\" (UID: \"9892a320-3e31-48df-bf8d-98441dd429ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.370349 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/35059f02-39e8-40b0-8295-8230591cd996-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zmdsl\" (UID: \"35059f02-39e8-40b0-8295-8230591cd996\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zmdsl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.370385 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vkxs\" (UniqueName: \"kubernetes.io/projected/86b3f820-9431-4cee-bcac-bb17c9a43fab-kube-api-access-5vkxs\") pod \"kube-storage-version-migrator-operator-b67b599dd-9bj8l\" (UID: \"86b3f820-9431-4cee-bcac-bb17c9a43fab\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.370464 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8jv9\" (UniqueName: \"kubernetes.io/projected/3369a23a-c296-4fda-8318-85dd8cea9913-kube-api-access-n8jv9\") pod \"dns-operator-744455d44c-9qdq5\" (UID: \"3369a23a-c296-4fda-8318-85dd8cea9913\") " pod="openshift-dns-operator/dns-operator-744455d44c-9qdq5" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.370485 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af533a06-13ef-4813-979d-875ffa3639dc-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8fgbp\" (UID: \"af533a06-13ef-4813-979d-875ffa3639dc\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.371496 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7510a625-89f5-43c1-9ea2-7b836c01c676-auth-proxy-config\") pod \"machine-config-operator-74547568cd-x2k8g\" (UID: \"7510a625-89f5-43c1-9ea2-7b836c01c676\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.372492 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kd6cl\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.378248 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.398367 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.418571 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.438635 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.445646 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/35059f02-39e8-40b0-8295-8230591cd996-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zmdsl\" (UID: \"35059f02-39e8-40b0-8295-8230591cd996\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zmdsl" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.458102 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.478918 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.485220 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/73fe25f4-c8e5-4d1e-a26e-25a3f0689593-profile-collector-cert\") pod \"olm-operator-6b444d44fb-dtp89\" (UID: \"73fe25f4-c8e5-4d1e-a26e-25a3f0689593\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.486820 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7ec3567f-1218-441b-b33d-054f0c6a2e2d-secret-volume\") pod \"collect-profiles-29400060-j98hw\" (UID: \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.499619 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.504524 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/404e867b-354a-45f7-b27a-10a7f3842555-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-chvs8\" (UID: \"404e867b-354a-45f7-b27a-10a7f3842555\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-chvs8" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.519665 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.539067 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.542280 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af533a06-13ef-4813-979d-875ffa3639dc-service-ca-bundle\") pod \"authentication-operator-69f744f599-8fgbp\" (UID: \"af533a06-13ef-4813-979d-875ffa3639dc\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.559557 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.566515 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af533a06-13ef-4813-979d-875ffa3639dc-serving-cert\") pod \"authentication-operator-69f744f599-8fgbp\" (UID: \"af533a06-13ef-4813-979d-875ffa3639dc\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.580545 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.599914 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.619516 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.621823 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af533a06-13ef-4813-979d-875ffa3639dc-config\") pod \"authentication-operator-69f744f599-8fgbp\" (UID: \"af533a06-13ef-4813-979d-875ffa3639dc\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.651112 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.653939 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af533a06-13ef-4813-979d-875ffa3639dc-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8fgbp\" (UID: \"af533a06-13ef-4813-979d-875ffa3639dc\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.659321 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.678484 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.699517 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.705508 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3369a23a-c296-4fda-8318-85dd8cea9913-metrics-tls\") pod \"dns-operator-744455d44c-9qdq5\" (UID: \"3369a23a-c296-4fda-8318-85dd8cea9913\") " pod="openshift-dns-operator/dns-operator-744455d44c-9qdq5" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.719447 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.739757 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.759155 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.778905 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.798621 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.818865 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.824455 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f140bda-5ace-46c7-a5a7-242b6bf58207-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5cdnv\" (UID: \"4f140bda-5ace-46c7-a5a7-242b6bf58207\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.838337 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.858551 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.861870 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f140bda-5ace-46c7-a5a7-242b6bf58207-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5cdnv\" (UID: \"4f140bda-5ace-46c7-a5a7-242b6bf58207\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.879534 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.899707 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.919864 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.939222 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.959760 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.961825 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7510a625-89f5-43c1-9ea2-7b836c01c676-images\") pod \"machine-config-operator-74547568cd-x2k8g\" (UID: \"7510a625-89f5-43c1-9ea2-7b836c01c676\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.979283 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 24 17:03:28 crc kubenswrapper[4777]: I1124 17:03:28.999378 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.019355 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.039531 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.045320 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7510a625-89f5-43c1-9ea2-7b836c01c676-proxy-tls\") pod \"machine-config-operator-74547568cd-x2k8g\" (UID: \"7510a625-89f5-43c1-9ea2-7b836c01c676\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.059513 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.078594 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.084937 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/73fe25f4-c8e5-4d1e-a26e-25a3f0689593-srv-cert\") pod \"olm-operator-6b444d44fb-dtp89\" (UID: \"73fe25f4-c8e5-4d1e-a26e-25a3f0689593\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.098610 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.117449 4777 request.go:700] Waited for 1.009468877s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-scheduler-operator/secrets?fieldSelector=metadata.name%3Dkube-scheduler-operator-serving-cert&limit=500&resourceVersion=0 Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.119113 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.125724 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9892a320-3e31-48df-bf8d-98441dd429ca-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncqcl\" (UID: \"9892a320-3e31-48df-bf8d-98441dd429ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.138258 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.142114 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9892a320-3e31-48df-bf8d-98441dd429ca-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncqcl\" (UID: \"9892a320-3e31-48df-bf8d-98441dd429ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.159547 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.181514 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.199600 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.205872 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/903c1f59-fc19-48a1-941e-d67c125e0d42-serving-cert\") pod \"service-ca-operator-777779d784-7dhx6\" (UID: \"903c1f59-fc19-48a1-941e-d67c125e0d42\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.218964 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.238960 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.241571 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/903c1f59-fc19-48a1-941e-d67c125e0d42-config\") pod \"service-ca-operator-777779d784-7dhx6\" (UID: \"903c1f59-fc19-48a1-941e-d67c125e0d42\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.260554 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.279416 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.299074 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.306756 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86b3f820-9431-4cee-bcac-bb17c9a43fab-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9bj8l\" (UID: \"86b3f820-9431-4cee-bcac-bb17c9a43fab\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.320273 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.322033 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86b3f820-9431-4cee-bcac-bb17c9a43fab-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9bj8l\" (UID: \"86b3f820-9431-4cee-bcac-bb17c9a43fab\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.339860 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.358912 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 24 17:03:29 crc kubenswrapper[4777]: E1124 17:03:29.371359 4777 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Nov 24 17:03:29 crc kubenswrapper[4777]: E1124 17:03:29.371433 4777 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Nov 24 17:03:29 crc kubenswrapper[4777]: E1124 17:03:29.371463 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/194ce833-4172-4abf-929c-2a9c296a9675-metrics-tls podName:194ce833-4172-4abf-929c-2a9c296a9675 nodeName:}" failed. No retries permitted until 2025-11-24 17:03:29.871430509 +0000 UTC m=+168.030625588 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/194ce833-4172-4abf-929c-2a9c296a9675-metrics-tls") pod "dns-default-wbx6n" (UID: "194ce833-4172-4abf-929c-2a9c296a9675") : failed to sync secret cache: timed out waiting for the condition Nov 24 17:03:29 crc kubenswrapper[4777]: E1124 17:03:29.371508 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/194ce833-4172-4abf-929c-2a9c296a9675-config-volume podName:194ce833-4172-4abf-929c-2a9c296a9675 nodeName:}" failed. No retries permitted until 2025-11-24 17:03:29.87148647 +0000 UTC m=+168.030681559 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/194ce833-4172-4abf-929c-2a9c296a9675-config-volume") pod "dns-default-wbx6n" (UID: "194ce833-4172-4abf-929c-2a9c296a9675") : failed to sync configmap cache: timed out waiting for the condition Nov 24 17:03:29 crc kubenswrapper[4777]: E1124 17:03:29.371701 4777 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Nov 24 17:03:29 crc kubenswrapper[4777]: E1124 17:03:29.371935 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7ec3567f-1218-441b-b33d-054f0c6a2e2d-config-volume podName:7ec3567f-1218-441b-b33d-054f0c6a2e2d nodeName:}" failed. No retries permitted until 2025-11-24 17:03:29.871902002 +0000 UTC m=+168.031097091 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/7ec3567f-1218-441b-b33d-054f0c6a2e2d-config-volume") pod "collect-profiles-29400060-j98hw" (UID: "7ec3567f-1218-441b-b33d-054f0c6a2e2d") : failed to sync configmap cache: timed out waiting for the condition Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.379844 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.399743 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.428053 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.439507 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.461653 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.479553 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.499721 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.519157 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.539894 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.560561 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.579505 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.599913 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.619549 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.640016 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.659496 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.680591 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.699155 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.718730 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.738877 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.787271 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4w4r\" (UniqueName: \"kubernetes.io/projected/55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4-kube-api-access-b4w4r\") pod \"apiserver-7bbb656c7d-xlq9c\" (UID: \"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.808134 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlqfr\" (UniqueName: \"kubernetes.io/projected/8062b545-bbdd-4a99-97c6-3682da1cb40f-kube-api-access-zlqfr\") pod \"apiserver-76f77b778f-qz2kt\" (UID: \"8062b545-bbdd-4a99-97c6-3682da1cb40f\") " pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.828699 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q49hn\" (UniqueName: \"kubernetes.io/projected/92efc884-606f-4fca-9410-4856b39a985d-kube-api-access-q49hn\") pod \"machine-api-operator-5694c8668f-xtvn9\" (UID: \"92efc884-606f-4fca-9410-4856b39a985d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.848362 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.850396 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq2xk\" (UniqueName: \"kubernetes.io/projected/5a08bdbf-d95e-4795-92be-f30739d00a6a-kube-api-access-sq2xk\") pod \"openshift-controller-manager-operator-756b6f6bc6-lhrpc\" (UID: \"5a08bdbf-d95e-4795-92be-f30739d00a6a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.860249 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.873542 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.880727 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.899505 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ec3567f-1218-441b-b33d-054f0c6a2e2d-config-volume\") pod \"collect-profiles-29400060-j98hw\" (UID: \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.899728 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/194ce833-4172-4abf-929c-2a9c296a9675-config-volume\") pod \"dns-default-wbx6n\" (UID: \"194ce833-4172-4abf-929c-2a9c296a9675\") " pod="openshift-dns/dns-default-wbx6n" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.899890 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/194ce833-4172-4abf-929c-2a9c296a9675-metrics-tls\") pod \"dns-default-wbx6n\" (UID: \"194ce833-4172-4abf-929c-2a9c296a9675\") " pod="openshift-dns/dns-default-wbx6n" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.901446 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ec3567f-1218-441b-b33d-054f0c6a2e2d-config-volume\") pod \"collect-profiles-29400060-j98hw\" (UID: \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.901677 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/194ce833-4172-4abf-929c-2a9c296a9675-config-volume\") pod \"dns-default-wbx6n\" (UID: \"194ce833-4172-4abf-929c-2a9c296a9675\") " pod="openshift-dns/dns-default-wbx6n" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.902002 4777 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.905244 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/194ce833-4172-4abf-929c-2a9c296a9675-metrics-tls\") pod \"dns-default-wbx6n\" (UID: \"194ce833-4172-4abf-929c-2a9c296a9675\") " pod="openshift-dns/dns-default-wbx6n" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.919905 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.940745 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.960249 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 24 17:03:29 crc kubenswrapper[4777]: I1124 17:03:29.980124 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.000517 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.024999 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.040477 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.072624 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.084002 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.115116 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpclm\" (UniqueName: \"kubernetes.io/projected/b8eb4c02-a9d2-416b-a363-da660f35d972-kube-api-access-zpclm\") pod \"service-ca-9c57cc56f-48b9w\" (UID: \"b8eb4c02-a9d2-416b-a363-da660f35d972\") " pod="openshift-service-ca/service-ca-9c57cc56f-48b9w" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.116214 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77rt5\" (UniqueName: \"kubernetes.io/projected/9773b6b0-ffe7-40ed-b714-c7266b46e298-kube-api-access-77rt5\") pod \"controller-manager-879f6c89f-kd6cl\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.120092 4777 request.go:700] Waited for 1.855130577s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/serviceaccounts/cluster-image-registry-operator/token Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.140300 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c"] Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.143717 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ff9e4751-a4e2-43c7-b75f-29ea3f7beca9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-s86qg\" (UID: \"ff9e4751-a4e2-43c7-b75f-29ea3f7beca9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.152867 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc"] Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.168486 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/493ed584-ded6-4027-ab8d-3946ecc089a4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6dsxq\" (UID: \"493ed584-ded6-4027-ab8d-3946ecc089a4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.182874 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttl95\" (UniqueName: \"kubernetes.io/projected/2ca6224b-5e61-422a-b961-5ceeea6bef3a-kube-api-access-ttl95\") pod \"console-operator-58897d9998-gcjhs\" (UID: \"2ca6224b-5e61-422a-b961-5ceeea6bef3a\") " pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.190439 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.197959 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vrg5\" (UniqueName: \"kubernetes.io/projected/eb260e29-56b9-4d87-9cee-7a8a3c336282-kube-api-access-7vrg5\") pod \"openshift-config-operator-7777fb866f-cw68j\" (UID: \"eb260e29-56b9-4d87-9cee-7a8a3c336282\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.223579 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-752sw\" (UniqueName: \"kubernetes.io/projected/10516d2d-106f-4155-ab2c-c0a302ca5cbf-kube-api-access-752sw\") pod \"console-f9d7485db-jxbj9\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.238074 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhv4w\" (UniqueName: \"kubernetes.io/projected/13c6ce51-e4c2-4397-8cd6-f101232d4bed-kube-api-access-zhv4w\") pod \"etcd-operator-b45778765-qp69m\" (UID: \"13c6ce51-e4c2-4397-8cd6-f101232d4bed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.253600 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8889\" (UniqueName: \"kubernetes.io/projected/0c79743f-3220-44ba-bc82-72ca21ee0daa-kube-api-access-m8889\") pod \"cluster-samples-operator-665b6dd947-6b6gc\" (UID: \"0c79743f-3220-44ba-bc82-72ca21ee0daa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6b6gc" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.259964 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6b6gc" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.267902 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" event={"ID":"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4","Type":"ContainerStarted","Data":"c0700ed91654f945f02f8dbbac3ee1b1ecd958a71302eb457251bf654be85049"} Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.269079 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc" event={"ID":"5a08bdbf-d95e-4795-92be-f30739d00a6a","Type":"ContainerStarted","Data":"947f3ec8ec84a7617628371c67c65fd68fa87f8c3d28f805c0392f72458cbf2d"} Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.275105 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cplxb\" (UniqueName: \"kubernetes.io/projected/90c90945-24a7-4f17-871c-7358792f0db5-kube-api-access-cplxb\") pod \"downloads-7954f5f757-vk8pz\" (UID: \"90c90945-24a7-4f17-871c-7358792f0db5\") " pod="openshift-console/downloads-7954f5f757-vk8pz" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.288176 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.294331 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.300062 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-xtvn9"] Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.300696 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-522qw\" (UniqueName: \"kubernetes.io/projected/5fdd1057-81bd-43f9-90f7-a1c33bfafba9-kube-api-access-522qw\") pod \"machine-approver-56656f9798-4gxrm\" (UID: \"5fdd1057-81bd-43f9-90f7-a1c33bfafba9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.303472 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.314378 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.318631 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktw2z\" (UniqueName: \"kubernetes.io/projected/0f99ce69-4367-49eb-84c8-534959fa1735-kube-api-access-ktw2z\") pod \"ingress-operator-5b745b69d9-d6bzx\" (UID: \"0f99ce69-4367-49eb-84c8-534959fa1735\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.321555 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq" Nov 24 17:03:30 crc kubenswrapper[4777]: W1124 17:03:30.329764 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92efc884_606f_4fca_9410_4856b39a985d.slice/crio-98247ba6492bf70e863094d63c4c9031e3a76c1aef9a730086d80a9efc18a0f1 WatchSource:0}: Error finding container 98247ba6492bf70e863094d63c4c9031e3a76c1aef9a730086d80a9efc18a0f1: Status 404 returned error can't find the container with id 98247ba6492bf70e863094d63c4c9031e3a76c1aef9a730086d80a9efc18a0f1 Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.334902 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkdnr\" (UniqueName: \"kubernetes.io/projected/2fd0dd98-6b46-4717-80c3-32d9502cae19-kube-api-access-wkdnr\") pod \"machine-config-controller-84d6567774-f6tp2\" (UID: \"2fd0dd98-6b46-4717-80c3-32d9502cae19\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.346499 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-48b9w" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.348260 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qz2kt"] Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.351407 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.359996 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.376508 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0f99ce69-4367-49eb-84c8-534959fa1735-bound-sa-token\") pod \"ingress-operator-5b745b69d9-d6bzx\" (UID: \"0f99ce69-4367-49eb-84c8-534959fa1735\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.385982 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6r9w\" (UniqueName: \"kubernetes.io/projected/2148eaac-ce88-4048-9ae9-9ff6df2d4747-kube-api-access-j6r9w\") pod \"route-controller-manager-6576b87f9c-zzd7t\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.400630 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sqqw\" (UniqueName: \"kubernetes.io/projected/54a60a6f-27d1-44d8-8372-512191c72f89-kube-api-access-8sqqw\") pod \"oauth-openshift-558db77b4-7hwq6\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.417256 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8h6s\" (UniqueName: \"kubernetes.io/projected/ff9e4751-a4e2-43c7-b75f-29ea3f7beca9-kube-api-access-c8h6s\") pod \"cluster-image-registry-operator-dc59b4c8b-s86qg\" (UID: \"ff9e4751-a4e2-43c7-b75f-29ea3f7beca9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.425793 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-gcjhs"] Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.437375 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nzmv\" (UniqueName: \"kubernetes.io/projected/698429dc-65d4-4b5f-b773-544f52c276c7-kube-api-access-5nzmv\") pod \"packageserver-d55dfcdfc-rrgd4\" (UID: \"698429dc-65d4-4b5f-b773-544f52c276c7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.454425 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hpzl\" (UniqueName: \"kubernetes.io/projected/4f140bda-5ace-46c7-a5a7-242b6bf58207-kube-api-access-2hpzl\") pod \"openshift-apiserver-operator-796bbdcf4f-5cdnv\" (UID: \"4f140bda-5ace-46c7-a5a7-242b6bf58207\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.468346 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.482162 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j4wk\" (UniqueName: \"kubernetes.io/projected/404e867b-354a-45f7-b27a-10a7f3842555-kube-api-access-7j4wk\") pod \"multus-admission-controller-857f4d67dd-chvs8\" (UID: \"404e867b-354a-45f7-b27a-10a7f3842555\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-chvs8" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.482554 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vk8pz" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.506464 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.514723 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfrbv\" (UniqueName: \"kubernetes.io/projected/903c1f59-fc19-48a1-941e-d67c125e0d42-kube-api-access-xfrbv\") pod \"service-ca-operator-777779d784-7dhx6\" (UID: \"903c1f59-fc19-48a1-941e-d67c125e0d42\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.520135 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.523846 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn9ck\" (UniqueName: \"kubernetes.io/projected/7ec3567f-1218-441b-b33d-054f0c6a2e2d-kube-api-access-qn9ck\") pod \"collect-profiles-29400060-j98hw\" (UID: \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.534614 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h689\" (UniqueName: \"kubernetes.io/projected/7510a625-89f5-43c1-9ea2-7b836c01c676-kube-api-access-4h689\") pod \"machine-config-operator-74547568cd-x2k8g\" (UID: \"7510a625-89f5-43c1-9ea2-7b836c01c676\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.550143 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.553173 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.554130 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9892a320-3e31-48df-bf8d-98441dd429ca-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncqcl\" (UID: \"9892a320-3e31-48df-bf8d-98441dd429ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.574244 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqqgh\" (UniqueName: \"kubernetes.io/projected/73fe25f4-c8e5-4d1e-a26e-25a3f0689593-kube-api-access-sqqgh\") pod \"olm-operator-6b444d44fb-dtp89\" (UID: \"73fe25f4-c8e5-4d1e-a26e-25a3f0689593\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.597911 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbvd5\" (UniqueName: \"kubernetes.io/projected/35059f02-39e8-40b0-8295-8230591cd996-kube-api-access-vbvd5\") pod \"control-plane-machine-set-operator-78cbb6b69f-zmdsl\" (UID: \"35059f02-39e8-40b0-8295-8230591cd996\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zmdsl" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.614657 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndw9c\" (UniqueName: \"kubernetes.io/projected/af533a06-13ef-4813-979d-875ffa3639dc-kube-api-access-ndw9c\") pod \"authentication-operator-69f744f599-8fgbp\" (UID: \"af533a06-13ef-4813-979d-875ffa3639dc\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.629715 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.636013 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.642872 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vcdh\" (UniqueName: \"kubernetes.io/projected/194ce833-4172-4abf-929c-2a9c296a9675-kube-api-access-7vcdh\") pod \"dns-default-wbx6n\" (UID: \"194ce833-4172-4abf-929c-2a9c296a9675\") " pod="openshift-dns/dns-default-wbx6n" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.673066 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vkxs\" (UniqueName: \"kubernetes.io/projected/86b3f820-9431-4cee-bcac-bb17c9a43fab-kube-api-access-5vkxs\") pod \"kube-storage-version-migrator-operator-b67b599dd-9bj8l\" (UID: \"86b3f820-9431-4cee-bcac-bb17c9a43fab\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.681538 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8jv9\" (UniqueName: \"kubernetes.io/projected/3369a23a-c296-4fda-8318-85dd8cea9913-kube-api-access-n8jv9\") pod \"dns-operator-744455d44c-9qdq5\" (UID: \"3369a23a-c296-4fda-8318-85dd8cea9913\") " pod="openshift-dns-operator/dns-operator-744455d44c-9qdq5" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.720724 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-registry-certificates\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.720780 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdjkt\" (UniqueName: \"kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-kube-api-access-wdjkt\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.720808 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6tcwx\" (UID: \"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab\") " pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.720844 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svd9m\" (UniqueName: \"kubernetes.io/projected/2ada3f56-91fc-4155-a12a-dfe0d898d282-kube-api-access-svd9m\") pod \"migrator-59844c95c7-82dhj\" (UID: \"2ada3f56-91fc-4155-a12a-dfe0d898d282\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-82dhj" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.720863 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6tcwx\" (UID: \"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab\") " pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.720885 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8cjs\" (UniqueName: \"kubernetes.io/projected/eacac718-a976-4e93-a7af-0414844177fb-kube-api-access-c8cjs\") pod \"router-default-5444994796-9cm8z\" (UID: \"eacac718-a976-4e93-a7af-0414844177fb\") " pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.720906 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrx47\" (UniqueName: \"kubernetes.io/projected/e0b913a5-16c4-4b8e-aa4e-0e04d9928537-kube-api-access-lrx47\") pod \"package-server-manager-789f6589d5-cqdzb\" (UID: \"e0b913a5-16c4-4b8e-aa4e-0e04d9928537\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.720923 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmh6j\" (UniqueName: \"kubernetes.io/projected/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-kube-api-access-zmh6j\") pod \"marketplace-operator-79b997595-6tcwx\" (UID: \"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab\") " pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.720941 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/eacac718-a976-4e93-a7af-0414844177fb-default-certificate\") pod \"router-default-5444994796-9cm8z\" (UID: \"eacac718-a976-4e93-a7af-0414844177fb\") " pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.720986 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-bound-sa-token\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.721008 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4c5288e3-5eeb-4107-8222-cd847fe099d5-profile-collector-cert\") pod \"catalog-operator-68c6474976-78g8k\" (UID: \"4c5288e3-5eeb-4107-8222-cd847fe099d5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.721031 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eacac718-a976-4e93-a7af-0414844177fb-metrics-certs\") pod \"router-default-5444994796-9cm8z\" (UID: \"eacac718-a976-4e93-a7af-0414844177fb\") " pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.721057 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.721097 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0b913a5-16c4-4b8e-aa4e-0e04d9928537-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-cqdzb\" (UID: \"e0b913a5-16c4-4b8e-aa4e-0e04d9928537\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.721141 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/eacac718-a976-4e93-a7af-0414844177fb-stats-auth\") pod \"router-default-5444994796-9cm8z\" (UID: \"eacac718-a976-4e93-a7af-0414844177fb\") " pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.721175 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eacac718-a976-4e93-a7af-0414844177fb-service-ca-bundle\") pod \"router-default-5444994796-9cm8z\" (UID: \"eacac718-a976-4e93-a7af-0414844177fb\") " pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.721197 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-trusted-ca\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.721220 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f56db4a4-25ec-4ebe-b290-e46d122a556c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-w57nx\" (UID: \"f56db4a4-25ec-4ebe-b290-e46d122a556c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.721242 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56db4a4-25ec-4ebe-b290-e46d122a556c-config\") pod \"kube-controller-manager-operator-78b949d7b-w57nx\" (UID: \"f56db4a4-25ec-4ebe-b290-e46d122a556c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.721294 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.721340 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l22s5\" (UniqueName: \"kubernetes.io/projected/4c5288e3-5eeb-4107-8222-cd847fe099d5-kube-api-access-l22s5\") pod \"catalog-operator-68c6474976-78g8k\" (UID: \"4c5288e3-5eeb-4107-8222-cd847fe099d5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.721381 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4c5288e3-5eeb-4107-8222-cd847fe099d5-srv-cert\") pod \"catalog-operator-68c6474976-78g8k\" (UID: \"4c5288e3-5eeb-4107-8222-cd847fe099d5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.721402 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f56db4a4-25ec-4ebe-b290-e46d122a556c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-w57nx\" (UID: \"f56db4a4-25ec-4ebe-b290-e46d122a556c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.721443 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-registry-tls\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.721463 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: E1124 17:03:30.722068 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:31.222050548 +0000 UTC m=+169.381245587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.724187 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zmdsl" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.738221 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-chvs8" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.748339 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.763418 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-9qdq5" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.790115 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.796940 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.814860 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.823550 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:30 crc kubenswrapper[4777]: E1124 17:03:30.823845 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:31.323823441 +0000 UTC m=+169.483018490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.823933 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0b913a5-16c4-4b8e-aa4e-0e04d9928537-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-cqdzb\" (UID: \"e0b913a5-16c4-4b8e-aa4e-0e04d9928537\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.824061 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/eacac718-a976-4e93-a7af-0414844177fb-stats-auth\") pod \"router-default-5444994796-9cm8z\" (UID: \"eacac718-a976-4e93-a7af-0414844177fb\") " pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.824115 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eacac718-a976-4e93-a7af-0414844177fb-service-ca-bundle\") pod \"router-default-5444994796-9cm8z\" (UID: \"eacac718-a976-4e93-a7af-0414844177fb\") " pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.824139 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-trusted-ca\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.824175 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f56db4a4-25ec-4ebe-b290-e46d122a556c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-w57nx\" (UID: \"f56db4a4-25ec-4ebe-b290-e46d122a556c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.824193 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56db4a4-25ec-4ebe-b290-e46d122a556c-config\") pod \"kube-controller-manager-operator-78b949d7b-w57nx\" (UID: \"f56db4a4-25ec-4ebe-b290-e46d122a556c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.824276 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ac194a53-ef69-4048-bc97-200861512652-cert\") pod \"ingress-canary-lzzxl\" (UID: \"ac194a53-ef69-4048-bc97-200861512652\") " pod="openshift-ingress-canary/ingress-canary-lzzxl" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.825101 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.825777 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eacac718-a976-4e93-a7af-0414844177fb-service-ca-bundle\") pod \"router-default-5444994796-9cm8z\" (UID: \"eacac718-a976-4e93-a7af-0414844177fb\") " pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.826402 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.826466 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-trusted-ca\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.826536 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rndln\" (UniqueName: \"kubernetes.io/projected/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-kube-api-access-rndln\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.826568 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l22s5\" (UniqueName: \"kubernetes.io/projected/4c5288e3-5eeb-4107-8222-cd847fe099d5-kube-api-access-l22s5\") pod \"catalog-operator-68c6474976-78g8k\" (UID: \"4c5288e3-5eeb-4107-8222-cd847fe099d5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.826796 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/81fa28f0-148e-4322-a5b7-947c18773c88-certs\") pod \"machine-config-server-bcj2h\" (UID: \"81fa28f0-148e-4322-a5b7-947c18773c88\") " pod="openshift-machine-config-operator/machine-config-server-bcj2h" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.827337 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56db4a4-25ec-4ebe-b290-e46d122a556c-config\") pod \"kube-controller-manager-operator-78b949d7b-w57nx\" (UID: \"f56db4a4-25ec-4ebe-b290-e46d122a556c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.827863 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4c5288e3-5eeb-4107-8222-cd847fe099d5-srv-cert\") pod \"catalog-operator-68c6474976-78g8k\" (UID: \"4c5288e3-5eeb-4107-8222-cd847fe099d5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.827954 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f56db4a4-25ec-4ebe-b290-e46d122a556c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-w57nx\" (UID: \"f56db4a4-25ec-4ebe-b290-e46d122a556c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.828183 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f56db4a4-25ec-4ebe-b290-e46d122a556c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-w57nx\" (UID: \"f56db4a4-25ec-4ebe-b290-e46d122a556c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.828194 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-socket-dir\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.828330 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.828407 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.828438 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-registry-tls\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.828545 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p89dh\" (UniqueName: \"kubernetes.io/projected/ac194a53-ef69-4048-bc97-200861512652-kube-api-access-p89dh\") pod \"ingress-canary-lzzxl\" (UID: \"ac194a53-ef69-4048-bc97-200861512652\") " pod="openshift-ingress-canary/ingress-canary-lzzxl" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.828606 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-registration-dir\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.828624 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-csi-data-dir\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.828659 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-registry-certificates\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.828694 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-mountpoint-dir\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.829204 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdjkt\" (UniqueName: \"kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-kube-api-access-wdjkt\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.829227 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6tcwx\" (UID: \"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab\") " pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.829274 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svd9m\" (UniqueName: \"kubernetes.io/projected/2ada3f56-91fc-4155-a12a-dfe0d898d282-kube-api-access-svd9m\") pod \"migrator-59844c95c7-82dhj\" (UID: \"2ada3f56-91fc-4155-a12a-dfe0d898d282\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-82dhj" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.829292 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/81fa28f0-148e-4322-a5b7-947c18773c88-node-bootstrap-token\") pod \"machine-config-server-bcj2h\" (UID: \"81fa28f0-148e-4322-a5b7-947c18773c88\") " pod="openshift-machine-config-operator/machine-config-server-bcj2h" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.829310 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6tcwx\" (UID: \"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab\") " pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.829325 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phps7\" (UniqueName: \"kubernetes.io/projected/81fa28f0-148e-4322-a5b7-947c18773c88-kube-api-access-phps7\") pod \"machine-config-server-bcj2h\" (UID: \"81fa28f0-148e-4322-a5b7-947c18773c88\") " pod="openshift-machine-config-operator/machine-config-server-bcj2h" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.829348 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8cjs\" (UniqueName: \"kubernetes.io/projected/eacac718-a976-4e93-a7af-0414844177fb-kube-api-access-c8cjs\") pod \"router-default-5444994796-9cm8z\" (UID: \"eacac718-a976-4e93-a7af-0414844177fb\") " pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.829364 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmh6j\" (UniqueName: \"kubernetes.io/projected/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-kube-api-access-zmh6j\") pod \"marketplace-operator-79b997595-6tcwx\" (UID: \"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab\") " pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.829379 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/eacac718-a976-4e93-a7af-0414844177fb-default-certificate\") pod \"router-default-5444994796-9cm8z\" (UID: \"eacac718-a976-4e93-a7af-0414844177fb\") " pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.829397 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-plugins-dir\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.829424 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-bound-sa-token\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.829442 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrx47\" (UniqueName: \"kubernetes.io/projected/e0b913a5-16c4-4b8e-aa4e-0e04d9928537-kube-api-access-lrx47\") pod \"package-server-manager-789f6589d5-cqdzb\" (UID: \"e0b913a5-16c4-4b8e-aa4e-0e04d9928537\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.829462 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4c5288e3-5eeb-4107-8222-cd847fe099d5-profile-collector-cert\") pod \"catalog-operator-68c6474976-78g8k\" (UID: \"4c5288e3-5eeb-4107-8222-cd847fe099d5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.829539 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.829555 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eacac718-a976-4e93-a7af-0414844177fb-metrics-certs\") pod \"router-default-5444994796-9cm8z\" (UID: \"eacac718-a976-4e93-a7af-0414844177fb\") " pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.833868 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.834225 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0b913a5-16c4-4b8e-aa4e-0e04d9928537-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-cqdzb\" (UID: \"e0b913a5-16c4-4b8e-aa4e-0e04d9928537\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.838714 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6tcwx\" (UID: \"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab\") " pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" Nov 24 17:03:30 crc kubenswrapper[4777]: E1124 17:03:30.841198 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:31.341179306 +0000 UTC m=+169.500374355 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.843165 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4c5288e3-5eeb-4107-8222-cd847fe099d5-srv-cert\") pod \"catalog-operator-68c6474976-78g8k\" (UID: \"4c5288e3-5eeb-4107-8222-cd847fe099d5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.843871 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/eacac718-a976-4e93-a7af-0414844177fb-stats-auth\") pod \"router-default-5444994796-9cm8z\" (UID: \"eacac718-a976-4e93-a7af-0414844177fb\") " pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.844139 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.845369 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eacac718-a976-4e93-a7af-0414844177fb-metrics-certs\") pod \"router-default-5444994796-9cm8z\" (UID: \"eacac718-a976-4e93-a7af-0414844177fb\") " pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.847574 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-registry-certificates\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.848290 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4c5288e3-5eeb-4107-8222-cd847fe099d5-profile-collector-cert\") pod \"catalog-operator-68c6474976-78g8k\" (UID: \"4c5288e3-5eeb-4107-8222-cd847fe099d5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.850912 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6tcwx\" (UID: \"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab\") " pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.855639 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/eacac718-a976-4e93-a7af-0414844177fb-default-certificate\") pod \"router-default-5444994796-9cm8z\" (UID: \"eacac718-a976-4e93-a7af-0414844177fb\") " pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.857266 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-registry-tls\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.857542 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-wbx6n" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.872042 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l22s5\" (UniqueName: \"kubernetes.io/projected/4c5288e3-5eeb-4107-8222-cd847fe099d5-kube-api-access-l22s5\") pod \"catalog-operator-68c6474976-78g8k\" (UID: \"4c5288e3-5eeb-4107-8222-cd847fe099d5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.921325 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svd9m\" (UniqueName: \"kubernetes.io/projected/2ada3f56-91fc-4155-a12a-dfe0d898d282-kube-api-access-svd9m\") pod \"migrator-59844c95c7-82dhj\" (UID: \"2ada3f56-91fc-4155-a12a-dfe0d898d282\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-82dhj" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.930096 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f56db4a4-25ec-4ebe-b290-e46d122a556c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-w57nx\" (UID: \"f56db4a4-25ec-4ebe-b290-e46d122a556c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.930679 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.931006 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ac194a53-ef69-4048-bc97-200861512652-cert\") pod \"ingress-canary-lzzxl\" (UID: \"ac194a53-ef69-4048-bc97-200861512652\") " pod="openshift-ingress-canary/ingress-canary-lzzxl" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.931086 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rndln\" (UniqueName: \"kubernetes.io/projected/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-kube-api-access-rndln\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.931111 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/81fa28f0-148e-4322-a5b7-947c18773c88-certs\") pod \"machine-config-server-bcj2h\" (UID: \"81fa28f0-148e-4322-a5b7-947c18773c88\") " pod="openshift-machine-config-operator/machine-config-server-bcj2h" Nov 24 17:03:30 crc kubenswrapper[4777]: E1124 17:03:30.931134 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:31.43111732 +0000 UTC m=+169.590312359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.931168 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs\") pod \"network-metrics-daemon-kxqzn\" (UID: \"960e1bed-abef-4f45-9475-37eee6bd780a\") " pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.931194 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-socket-dir\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.931244 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p89dh\" (UniqueName: \"kubernetes.io/projected/ac194a53-ef69-4048-bc97-200861512652-kube-api-access-p89dh\") pod \"ingress-canary-lzzxl\" (UID: \"ac194a53-ef69-4048-bc97-200861512652\") " pod="openshift-ingress-canary/ingress-canary-lzzxl" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.931269 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-registration-dir\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.931310 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-csi-data-dir\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.931337 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-mountpoint-dir\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.931415 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/81fa28f0-148e-4322-a5b7-947c18773c88-node-bootstrap-token\") pod \"machine-config-server-bcj2h\" (UID: \"81fa28f0-148e-4322-a5b7-947c18773c88\") " pod="openshift-machine-config-operator/machine-config-server-bcj2h" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.931440 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phps7\" (UniqueName: \"kubernetes.io/projected/81fa28f0-148e-4322-a5b7-947c18773c88-kube-api-access-phps7\") pod \"machine-config-server-bcj2h\" (UID: \"81fa28f0-148e-4322-a5b7-947c18773c88\") " pod="openshift-machine-config-operator/machine-config-server-bcj2h" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.931516 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-plugins-dir\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.931571 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.931951 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-mountpoint-dir\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: E1124 17:03:30.931990 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:31.431976233 +0000 UTC m=+169.591171282 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.932144 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-csi-data-dir\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.932242 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-socket-dir\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.932240 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-plugins-dir\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.932282 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-registration-dir\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.936433 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/960e1bed-abef-4f45-9475-37eee6bd780a-metrics-certs\") pod \"network-metrics-daemon-kxqzn\" (UID: \"960e1bed-abef-4f45-9475-37eee6bd780a\") " pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.936778 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/81fa28f0-148e-4322-a5b7-947c18773c88-certs\") pod \"machine-config-server-bcj2h\" (UID: \"81fa28f0-148e-4322-a5b7-947c18773c88\") " pod="openshift-machine-config-operator/machine-config-server-bcj2h" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.937402 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8cjs\" (UniqueName: \"kubernetes.io/projected/eacac718-a976-4e93-a7af-0414844177fb-kube-api-access-c8cjs\") pod \"router-default-5444994796-9cm8z\" (UID: \"eacac718-a976-4e93-a7af-0414844177fb\") " pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.937759 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/81fa28f0-148e-4322-a5b7-947c18773c88-node-bootstrap-token\") pod \"machine-config-server-bcj2h\" (UID: \"81fa28f0-148e-4322-a5b7-947c18773c88\") " pod="openshift-machine-config-operator/machine-config-server-bcj2h" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.938490 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ac194a53-ef69-4048-bc97-200861512652-cert\") pod \"ingress-canary-lzzxl\" (UID: \"ac194a53-ef69-4048-bc97-200861512652\") " pod="openshift-ingress-canary/ingress-canary-lzzxl" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.959287 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmh6j\" (UniqueName: \"kubernetes.io/projected/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-kube-api-access-zmh6j\") pod \"marketplace-operator-79b997595-6tcwx\" (UID: \"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab\") " pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.959607 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kxqzn" Nov 24 17:03:30 crc kubenswrapper[4777]: I1124 17:03:30.977716 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-bound-sa-token\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.002845 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrx47\" (UniqueName: \"kubernetes.io/projected/e0b913a5-16c4-4b8e-aa4e-0e04d9928537-kube-api-access-lrx47\") pod \"package-server-manager-789f6589d5-cqdzb\" (UID: \"e0b913a5-16c4-4b8e-aa4e-0e04d9928537\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.024018 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdjkt\" (UniqueName: \"kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-kube-api-access-wdjkt\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.032519 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:31 crc kubenswrapper[4777]: E1124 17:03:31.032850 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:31.532834542 +0000 UTC m=+169.692029591 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.033672 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.064561 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p89dh\" (UniqueName: \"kubernetes.io/projected/ac194a53-ef69-4048-bc97-200861512652-kube-api-access-p89dh\") pod \"ingress-canary-lzzxl\" (UID: \"ac194a53-ef69-4048-bc97-200861512652\") " pod="openshift-ingress-canary/ingress-canary-lzzxl" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.074994 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.081262 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-82dhj" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.083404 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rndln\" (UniqueName: \"kubernetes.io/projected/4ef71d5e-7018-4d3b-9eb1-abba188e4a39-kube-api-access-rndln\") pod \"csi-hostpathplugin-fsbnq\" (UID: \"4ef71d5e-7018-4d3b-9eb1-abba188e4a39\") " pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.093635 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phps7\" (UniqueName: \"kubernetes.io/projected/81fa28f0-148e-4322-a5b7-947c18773c88-kube-api-access-phps7\") pod \"machine-config-server-bcj2h\" (UID: \"81fa28f0-148e-4322-a5b7-947c18773c88\") " pod="openshift-machine-config-operator/machine-config-server-bcj2h" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.125335 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.137171 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.141554 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:31 crc kubenswrapper[4777]: E1124 17:03:31.145467 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:31.645445478 +0000 UTC m=+169.804640527 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.160167 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6b6gc"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.171689 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-jxbj9"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.173102 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-48b9w"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.173360 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.174025 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cw68j"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.176381 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qp69m"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.182555 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bcj2h" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.190444 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-lzzxl" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.246533 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:31 crc kubenswrapper[4777]: E1124 17:03:31.247595 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:31.747573642 +0000 UTC m=+169.906768691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.271140 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.296836 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc" event={"ID":"5a08bdbf-d95e-4795-92be-f30739d00a6a","Type":"ContainerStarted","Data":"930161602abe0be4272dda7bcd92b33b9c2ed7d0eda6e778bbc86755c19d863a"} Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.305544 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" event={"ID":"5fdd1057-81bd-43f9-90f7-a1c33bfafba9","Type":"ContainerStarted","Data":"c1392cc584eea13994368ee8f43d645bfdf44f5a6f2a1f30d15a2cdf1c09a81b"} Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.305585 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" event={"ID":"5fdd1057-81bd-43f9-90f7-a1c33bfafba9","Type":"ContainerStarted","Data":"839259df3a04e6a23e80c24c374d809412c36e761de0cfba6d7b4c82e9c82b78"} Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.307782 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" event={"ID":"13c6ce51-e4c2-4397-8cd6-f101232d4bed","Type":"ContainerStarted","Data":"c75ba02b11360666ccf24ad57cc6e04ac2d6700f5499cbd8a80350a72326f6d5"} Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.314072 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-gcjhs" event={"ID":"2ca6224b-5e61-422a-b961-5ceeea6bef3a","Type":"ContainerStarted","Data":"08b14a0a9707291bba0d58c31524c444cb7f6395200ddfcd310d737e857cda56"} Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.314122 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-gcjhs" event={"ID":"2ca6224b-5e61-422a-b961-5ceeea6bef3a","Type":"ContainerStarted","Data":"3824b6ac764a12bd7c9cc300bf4d35d80524b2b0c198ee0280d426cdae5e3a44"} Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.314255 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.316541 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" event={"ID":"eb260e29-56b9-4d87-9cee-7a8a3c336282","Type":"ContainerStarted","Data":"d6cd29e2260c33cfe15cb9008db47dda06d667afaf3fb62624b1ed8926371c06"} Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.317027 4777 patch_prober.go:28] interesting pod/console-operator-58897d9998-gcjhs container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.317072 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-gcjhs" podUID="2ca6224b-5e61-422a-b961-5ceeea6bef3a" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.321005 4777 generic.go:334] "Generic (PLEG): container finished" podID="55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4" containerID="45c07824ce713aeaecea1e0b9bffb46e271175d22b6f96ed9ec39d600eee36c7" exitCode=0 Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.321107 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" event={"ID":"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4","Type":"ContainerDied","Data":"45c07824ce713aeaecea1e0b9bffb46e271175d22b6f96ed9ec39d600eee36c7"} Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.327598 4777 generic.go:334] "Generic (PLEG): container finished" podID="8062b545-bbdd-4a99-97c6-3682da1cb40f" containerID="be522548c22b3df7a2f4bfc59c59486c8afa16cedbd51e574967aa7f0458fe5a" exitCode=0 Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.327684 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" event={"ID":"8062b545-bbdd-4a99-97c6-3682da1cb40f","Type":"ContainerDied","Data":"be522548c22b3df7a2f4bfc59c59486c8afa16cedbd51e574967aa7f0458fe5a"} Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.327723 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" event={"ID":"8062b545-bbdd-4a99-97c6-3682da1cb40f","Type":"ContainerStarted","Data":"ea1b5b3fe602225938b22ca339f2390c30e01583b42226b9dce0f7c75eb40f42"} Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.329514 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9cm8z" event={"ID":"eacac718-a976-4e93-a7af-0414844177fb","Type":"ContainerStarted","Data":"661f0d9768589f4956710373a806bcb80b6ac2e593bee3698aadd29618590bcc"} Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.331139 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" event={"ID":"92efc884-606f-4fca-9410-4856b39a985d","Type":"ContainerStarted","Data":"e0d27882f9ef8d9173b82bc5ee33d48db213f828914b2de7fdc2d3b2fbf13cc4"} Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.331163 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" event={"ID":"92efc884-606f-4fca-9410-4856b39a985d","Type":"ContainerStarted","Data":"10ce195cb2710325221903cc6006e12318d11800eb463d2168cecc6affffd4f2"} Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.331175 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" event={"ID":"92efc884-606f-4fca-9410-4856b39a985d","Type":"ContainerStarted","Data":"98247ba6492bf70e863094d63c4c9031e3a76c1aef9a730086d80a9efc18a0f1"} Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.349113 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:31 crc kubenswrapper[4777]: E1124 17:03:31.350147 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:31.850136048 +0000 UTC m=+170.009331097 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.451154 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:31 crc kubenswrapper[4777]: E1124 17:03:31.452595 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:31.95258006 +0000 UTC m=+170.111775109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.555468 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:31 crc kubenswrapper[4777]: E1124 17:03:31.556511 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:32.056490364 +0000 UTC m=+170.215685413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.587545 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kd6cl"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.627291 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.630311 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.636955 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.645715 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.651940 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vk8pz"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.664007 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.666235 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:31 crc kubenswrapper[4777]: E1124 17:03:31.666688 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:32.166666963 +0000 UTC m=+170.325862012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.667274 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.677094 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7hwq6"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.678741 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw"] Nov 24 17:03:31 crc kubenswrapper[4777]: W1124 17:03:31.701688 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod698429dc_65d4_4b5f_b773_544f52c276c7.slice/crio-26c5f2e454c135bb24628446d038ba680350d02727b3aa759465a04072d6b9c7 WatchSource:0}: Error finding container 26c5f2e454c135bb24628446d038ba680350d02727b3aa759465a04072d6b9c7: Status 404 returned error can't find the container with id 26c5f2e454c135bb24628446d038ba680350d02727b3aa759465a04072d6b9c7 Nov 24 17:03:31 crc kubenswrapper[4777]: W1124 17:03:31.731936 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54a60a6f_27d1_44d8_8372_512191c72f89.slice/crio-6ab0e58e616b76783bdbdb740d6a80a7ed5c41de305fa1e108f581c3c53349ea WatchSource:0}: Error finding container 6ab0e58e616b76783bdbdb740d6a80a7ed5c41de305fa1e108f581c3c53349ea: Status 404 returned error can't find the container with id 6ab0e58e616b76783bdbdb740d6a80a7ed5c41de305fa1e108f581c3c53349ea Nov 24 17:03:31 crc kubenswrapper[4777]: W1124 17:03:31.733804 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2148eaac_ce88_4048_9ae9_9ff6df2d4747.slice/crio-fe754e761e4badd714b65f963f8488c330a3b82534ca3aab6610b311eb7192b5 WatchSource:0}: Error finding container fe754e761e4badd714b65f963f8488c330a3b82534ca3aab6610b311eb7192b5: Status 404 returned error can't find the container with id fe754e761e4badd714b65f963f8488c330a3b82534ca3aab6610b311eb7192b5 Nov 24 17:03:31 crc kubenswrapper[4777]: W1124 17:03:31.735018 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ec3567f_1218_441b_b33d_054f0c6a2e2d.slice/crio-d48f19487aea555851dd6772c9a9157e6bd3f6e58ec6be8abc03d42c6a184797 WatchSource:0}: Error finding container d48f19487aea555851dd6772c9a9157e6bd3f6e58ec6be8abc03d42c6a184797: Status 404 returned error can't find the container with id d48f19487aea555851dd6772c9a9157e6bd3f6e58ec6be8abc03d42c6a184797 Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.767546 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:31 crc kubenswrapper[4777]: E1124 17:03:31.768011 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:32.267994144 +0000 UTC m=+170.427189193 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.868247 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:31 crc kubenswrapper[4777]: E1124 17:03:31.868690 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:32.368674967 +0000 UTC m=+170.527870016 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.930347 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.933496 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.940165 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.943046 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.959903 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kxqzn"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.961748 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zmdsl"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.972103 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:31 crc kubenswrapper[4777]: E1124 17:03:31.972518 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:32.472502738 +0000 UTC m=+170.631697807 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.983543 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9qdq5"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.988768 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-chvs8"] Nov 24 17:03:31 crc kubenswrapper[4777]: I1124 17:03:31.992769 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6"] Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.012456 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-wbx6n"] Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.077568 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:32 crc kubenswrapper[4777]: E1124 17:03:32.088790 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:32.588764037 +0000 UTC m=+170.747959086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.116209 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-82dhj"] Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.126647 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8fgbp"] Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.129072 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-fsbnq"] Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.154128 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89"] Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.177068 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k"] Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.179612 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx"] Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.181099 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb"] Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.187316 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:32 crc kubenswrapper[4777]: E1124 17:03:32.187638 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:32.68762549 +0000 UTC m=+170.846820539 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.188842 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-gcjhs" podStartSLOduration=145.188820343 podStartE2EDuration="2m25.188820343s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:32.167441626 +0000 UTC m=+170.326636675" watchObservedRunningTime="2025-11-24 17:03:32.188820343 +0000 UTC m=+170.348015392" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.236874 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6tcwx"] Nov 24 17:03:32 crc kubenswrapper[4777]: W1124 17:03:32.285071 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ef71d5e_7018_4d3b_9eb1_abba188e4a39.slice/crio-d3268a53422ada45616bedfa435fa5f754eb6f4ff5c14a62ad572f6ae072f083 WatchSource:0}: Error finding container d3268a53422ada45616bedfa435fa5f754eb6f4ff5c14a62ad572f6ae072f083: Status 404 returned error can't find the container with id d3268a53422ada45616bedfa435fa5f754eb6f4ff5c14a62ad572f6ae072f083 Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.289814 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:32 crc kubenswrapper[4777]: E1124 17:03:32.289999 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:32.789932648 +0000 UTC m=+170.949127697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.290755 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:32 crc kubenswrapper[4777]: W1124 17:03:32.291479 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73fe25f4_c8e5_4d1e_a26e_25a3f0689593.slice/crio-3e0880c9578349830e0de8b38a8c934bf055ab461286962c2d33200caa964847 WatchSource:0}: Error finding container 3e0880c9578349830e0de8b38a8c934bf055ab461286962c2d33200caa964847: Status 404 returned error can't find the container with id 3e0880c9578349830e0de8b38a8c934bf055ab461286962c2d33200caa964847 Nov 24 17:03:32 crc kubenswrapper[4777]: E1124 17:03:32.291527 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:32.791467211 +0000 UTC m=+170.950662250 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.295170 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-lzzxl"] Nov 24 17:03:32 crc kubenswrapper[4777]: W1124 17:03:32.318076 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0b913a5_16c4_4b8e_aa4e_0e04d9928537.slice/crio-ccf4846963c8e6f837bef5801f810f1af1aa5fb5a25f6c5b3695c9718f552e44 WatchSource:0}: Error finding container ccf4846963c8e6f837bef5801f810f1af1aa5fb5a25f6c5b3695c9718f552e44: Status 404 returned error can't find the container with id ccf4846963c8e6f837bef5801f810f1af1aa5fb5a25f6c5b3695c9718f552e44 Nov 24 17:03:32 crc kubenswrapper[4777]: W1124 17:03:32.329876 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c5288e3_5eeb_4107_8222_cd847fe099d5.slice/crio-9cc4f224c0815c739d8c0909ab7af08461bebb42f0bc8da2c878061ec30fe149 WatchSource:0}: Error finding container 9cc4f224c0815c739d8c0909ab7af08461bebb42f0bc8da2c878061ec30fe149: Status 404 returned error can't find the container with id 9cc4f224c0815c739d8c0909ab7af08461bebb42f0bc8da2c878061ec30fe149 Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.347314 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" event={"ID":"7ec3567f-1218-441b-b33d-054f0c6a2e2d","Type":"ContainerStarted","Data":"d48f19487aea555851dd6772c9a9157e6bd3f6e58ec6be8abc03d42c6a184797"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.361011 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6b6gc" event={"ID":"0c79743f-3220-44ba-bc82-72ca21ee0daa","Type":"ContainerStarted","Data":"3c964a3b48c97b05cd0711dd51a2756179f90d04ee132bbc4aac164705b94851"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.361082 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6b6gc" event={"ID":"0c79743f-3220-44ba-bc82-72ca21ee0daa","Type":"ContainerStarted","Data":"1bd391a18b3fbe75fb428463f8840e5f2ad5692607ce1fb53749abbfeec46caf"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.379402 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv" event={"ID":"4f140bda-5ace-46c7-a5a7-242b6bf58207","Type":"ContainerStarted","Data":"033b9a0274d8f6e31d6165d4bfa1398e08d63471239127128a069d91468abef9"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.386026 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wbx6n" event={"ID":"194ce833-4172-4abf-929c-2a9c296a9675","Type":"ContainerStarted","Data":"32f7b72d98fe55885060c000ec4e08b18bf21f80c5ca39d591b064d7088d75e5"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.389440 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" event={"ID":"2148eaac-ce88-4048-9ae9-9ff6df2d4747","Type":"ContainerStarted","Data":"f3f60291ec83f38106556bbf549de51d0ce0ff286edbec947edbc1a5d4c3fa66"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.389728 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.389753 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" event={"ID":"2148eaac-ce88-4048-9ae9-9ff6df2d4747","Type":"ContainerStarted","Data":"fe754e761e4badd714b65f963f8488c330a3b82534ca3aab6610b311eb7192b5"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.392504 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:32 crc kubenswrapper[4777]: E1124 17:03:32.392803 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:32.892784432 +0000 UTC m=+171.051979481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.394343 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9qdq5" event={"ID":"3369a23a-c296-4fda-8318-85dd8cea9913","Type":"ContainerStarted","Data":"e551f21dba2a08a0b60fbd3996542b3a9728fb8050a232f36c638f481bcf860d"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.397356 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6" event={"ID":"903c1f59-fc19-48a1-941e-d67c125e0d42","Type":"ContainerStarted","Data":"f35f55d83f9b8e63a6eb6759aff485872aa647be4d95ab73f3aaa552edc9ae51"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.402700 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" event={"ID":"af533a06-13ef-4813-979d-875ffa3639dc","Type":"ContainerStarted","Data":"013d6cacde84f9aa7ba9a48a7147b54f0990463f27d69e1d81e2474e2f497638"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.402825 4777 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-zzd7t container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.403040 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" podUID="2148eaac-ce88-4048-9ae9-9ff6df2d4747" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.424156 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" event={"ID":"8062b545-bbdd-4a99-97c6-3682da1cb40f","Type":"ContainerStarted","Data":"b50577e488cf532712bf6314430a2463f09ab194f20043ee7cfc1d8b0dc3e675"} Nov 24 17:03:32 crc kubenswrapper[4777]: W1124 17:03:32.430872 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac194a53_ef69_4048_bc97_200861512652.slice/crio-66ae89cd2672a924796d425d0489baf595020dfb8835e858e49fa4fa9176d739 WatchSource:0}: Error finding container 66ae89cd2672a924796d425d0489baf595020dfb8835e858e49fa4fa9176d739: Status 404 returned error can't find the container with id 66ae89cd2672a924796d425d0489baf595020dfb8835e858e49fa4fa9176d739 Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.436714 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" event={"ID":"7510a625-89f5-43c1-9ea2-7b836c01c676","Type":"ContainerStarted","Data":"a52b793c44b1751269a2ff5a897dff6d77b20ad7b6fc47a6fff48b2120c42845"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.447191 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq" event={"ID":"493ed584-ded6-4027-ab8d-3946ecc089a4","Type":"ContainerStarted","Data":"6f578deae8c4c0d512132afb44ffd13df7b456bc2fc95731006c2c086ff11a22"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.460247 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2" event={"ID":"2fd0dd98-6b46-4717-80c3-32d9502cae19","Type":"ContainerStarted","Data":"1b05d405845386872202775018447a275a975b6f96607c123e91a19d3198e7bf"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.466844 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9cm8z" event={"ID":"eacac718-a976-4e93-a7af-0414844177fb","Type":"ContainerStarted","Data":"0d0fb923daf42a0c7ce5d4ea979c283d8d9c1ad9eb080ddc9d2da6bda638fed1"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.468585 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zmdsl" event={"ID":"35059f02-39e8-40b0-8295-8230591cd996","Type":"ContainerStarted","Data":"0b4f0629a20921a6bec7941887dea735e497bc75c7221d6d2f65d093dde24878"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.497345 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-chvs8" event={"ID":"404e867b-354a-45f7-b27a-10a7f3842555","Type":"ContainerStarted","Data":"cfe9d3b7a5a5fcf56df15d5d97bb350140a4a7986879478fac4d65df239e2ab9"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.498012 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:32 crc kubenswrapper[4777]: E1124 17:03:32.499349 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:32.99933877 +0000 UTC m=+171.158533819 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.518818 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-48b9w" event={"ID":"b8eb4c02-a9d2-416b-a363-da660f35d972","Type":"ContainerStarted","Data":"d2fcf8b75bb75cffb86684a30630b0a724ee3e6a289e2d575f8bb1ceee38675f"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.518869 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-48b9w" event={"ID":"b8eb4c02-a9d2-416b-a363-da660f35d972","Type":"ContainerStarted","Data":"0b47062d48cb10d2d1afc9a9abb5cc6000d7c108d9cfcf321b243f47bf3da83b"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.529176 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lhrpc" podStartSLOduration=145.529158533 podStartE2EDuration="2m25.529158533s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:32.527492246 +0000 UTC m=+170.686687295" watchObservedRunningTime="2025-11-24 17:03:32.529158533 +0000 UTC m=+170.688353582" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.537343 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" event={"ID":"960e1bed-abef-4f45-9475-37eee6bd780a","Type":"ContainerStarted","Data":"3cde3b2f0ca699683eda7cffcc32217cf1043d95006725183f1540f1e0e3f9d8"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.549255 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" event={"ID":"0f99ce69-4367-49eb-84c8-534959fa1735","Type":"ContainerStarted","Data":"cb2a22a83b5b0a7c06c943c697c0dcf9253df2fcca78c851c1c314c44b6c4db9"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.549291 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" event={"ID":"0f99ce69-4367-49eb-84c8-534959fa1735","Type":"ContainerStarted","Data":"cf0eec64ed883880e4fc783795a60878995aa185d404c2e3e6355af3fffe7f3f"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.552949 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" event={"ID":"73fe25f4-c8e5-4d1e-a26e-25a3f0689593","Type":"ContainerStarted","Data":"3e0880c9578349830e0de8b38a8c934bf055ab461286962c2d33200caa964847"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.565320 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-jxbj9" event={"ID":"10516d2d-106f-4155-ab2c-c0a302ca5cbf","Type":"ContainerStarted","Data":"102ae36ed9ce8ddf96acc06d1168c28ad8ba207eb6d34f0b39e8483ebc33be32"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.565363 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-jxbj9" event={"ID":"10516d2d-106f-4155-ab2c-c0a302ca5cbf","Type":"ContainerStarted","Data":"01960da356fb01e61c8a30a8cf978e1ce71ff26983c4ad232650a6b79a93c3e1"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.571414 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" event={"ID":"698429dc-65d4-4b5f-b773-544f52c276c7","Type":"ContainerStarted","Data":"3cfa3ab8eeb77589c9df50601c17206954205ddb615f9209402e2f7c43bcb19f"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.571456 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" event={"ID":"698429dc-65d4-4b5f-b773-544f52c276c7","Type":"ContainerStarted","Data":"26c5f2e454c135bb24628446d038ba680350d02727b3aa759465a04072d6b9c7"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.571742 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.573924 4777 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-rrgd4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" start-of-body= Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.574011 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" podUID="698429dc-65d4-4b5f-b773-544f52c276c7" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.578855 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx" event={"ID":"f56db4a4-25ec-4ebe-b290-e46d122a556c","Type":"ContainerStarted","Data":"bb9bb6d3402aeb1301955bb26a2476f18ca016713660b7531355cd849ed3c0ee"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.587307 4777 generic.go:334] "Generic (PLEG): container finished" podID="eb260e29-56b9-4d87-9cee-7a8a3c336282" containerID="abd7d872a903706fe7fa493e61d52687f001c45377c5a89acdde9688e9d823ff" exitCode=0 Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.588083 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" event={"ID":"eb260e29-56b9-4d87-9cee-7a8a3c336282","Type":"ContainerDied","Data":"abd7d872a903706fe7fa493e61d52687f001c45377c5a89acdde9688e9d823ff"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.593598 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-xtvn9" podStartSLOduration=145.593581643 podStartE2EDuration="2m25.593581643s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:32.591726771 +0000 UTC m=+170.750921820" watchObservedRunningTime="2025-11-24 17:03:32.593581643 +0000 UTC m=+170.752776702" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.597109 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" event={"ID":"ff9e4751-a4e2-43c7-b75f-29ea3f7beca9","Type":"ContainerStarted","Data":"a105671c2c8bb1d831bef0c197f2eeb6aee62a11078a1b749ad6605140389435"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.599466 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.599579 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" event={"ID":"4c5288e3-5eeb-4107-8222-cd847fe099d5","Type":"ContainerStarted","Data":"9cc4f224c0815c739d8c0909ab7af08461bebb42f0bc8da2c878061ec30fe149"} Nov 24 17:03:32 crc kubenswrapper[4777]: E1124 17:03:32.600286 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:33.10027063 +0000 UTC m=+171.259465679 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.610353 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-82dhj" event={"ID":"2ada3f56-91fc-4155-a12a-dfe0d898d282","Type":"ContainerStarted","Data":"4454699ec347b4afa4ba375b478e5d7fc5345cf5d9540a91d4187eef3bf3033e"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.657716 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" event={"ID":"54a60a6f-27d1-44d8-8372-512191c72f89","Type":"ContainerStarted","Data":"6ab0e58e616b76783bdbdb740d6a80a7ed5c41de305fa1e108f581c3c53349ea"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.686513 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl" event={"ID":"9892a320-3e31-48df-bf8d-98441dd429ca","Type":"ContainerStarted","Data":"b4329e496b745c3d730b1750c123187be5cc0db36be223b32b4b45d58ce1942f"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.690655 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" event={"ID":"5fdd1057-81bd-43f9-90f7-a1c33bfafba9","Type":"ContainerStarted","Data":"63f27c3d81c3074b5dc223e38495a5ea8e9e8a5589b86f6a9dc70acbec9c13b9"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.701714 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:32 crc kubenswrapper[4777]: E1124 17:03:32.704311 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:33.204298477 +0000 UTC m=+171.363493526 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.725551 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" event={"ID":"13c6ce51-e4c2-4397-8cd6-f101232d4bed","Type":"ContainerStarted","Data":"b0d38789bf3f69475ced4d91053b083a1122070f547f1fee91802f24fb2249d8"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.754089 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l" event={"ID":"86b3f820-9431-4cee-bcac-bb17c9a43fab","Type":"ContainerStarted","Data":"836ed9fcce97a89d5017e9153587c30bd2cd34e96cbbb58a515757055602845b"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.758354 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" podStartSLOduration=145.758333986 podStartE2EDuration="2m25.758333986s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:32.754119029 +0000 UTC m=+170.913314078" watchObservedRunningTime="2025-11-24 17:03:32.758333986 +0000 UTC m=+170.917529035" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.790265 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bcj2h" event={"ID":"81fa28f0-148e-4322-a5b7-947c18773c88","Type":"ContainerStarted","Data":"08b6bc9750ccfed9c3cf6c1210dddefedcb790f84613a5bb9771e41c6b399535"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.790412 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bcj2h" event={"ID":"81fa28f0-148e-4322-a5b7-947c18773c88","Type":"ContainerStarted","Data":"cd074551ba081d2b8df0d6d9899bb744dffb3ae79dac04f44c9f44d29d093aa3"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.807883 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:32 crc kubenswrapper[4777]: E1124 17:03:32.813822 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:33.313799576 +0000 UTC m=+171.472994625 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.820472 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vk8pz" event={"ID":"90c90945-24a7-4f17-871c-7358792f0db5","Type":"ContainerStarted","Data":"3d2222df8f89be3b3b5c6d69c3d52ab7fe55b885fefa40c92f83fe5702297f57"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.821163 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vk8pz" event={"ID":"90c90945-24a7-4f17-871c-7358792f0db5","Type":"ContainerStarted","Data":"5c610109544f63ecfafa24f114a5da5a2ac109360508d6b08ebe19235af45f45"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.821268 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-vk8pz" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.836675 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" event={"ID":"9773b6b0-ffe7-40ed-b714-c7266b46e298","Type":"ContainerStarted","Data":"b4d91654bd09070f705540d28bab0d20b5c67cdfc36852d890fe776f6a4340d4"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.837883 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.853902 4777 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk8pz container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.853978 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vk8pz" podUID="90c90945-24a7-4f17-871c-7358792f0db5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.854280 4777 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-kd6cl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.854302 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" podUID="9773b6b0-ffe7-40ed-b714-c7266b46e298" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.862242 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" event={"ID":"4ef71d5e-7018-4d3b-9eb1-abba188e4a39","Type":"ContainerStarted","Data":"d3268a53422ada45616bedfa435fa5f754eb6f4ff5c14a62ad572f6ae072f083"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.865310 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb" event={"ID":"e0b913a5-16c4-4b8e-aa4e-0e04d9928537","Type":"ContainerStarted","Data":"ccf4846963c8e6f837bef5801f810f1af1aa5fb5a25f6c5b3695c9718f552e44"} Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.871077 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-gcjhs" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.872773 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-jxbj9" podStartSLOduration=145.872758004 podStartE2EDuration="2m25.872758004s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:32.872306881 +0000 UTC m=+171.031501930" watchObservedRunningTime="2025-11-24 17:03:32.872758004 +0000 UTC m=+171.031953043" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.874402 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4gxrm" podStartSLOduration=145.874396329 podStartE2EDuration="2m25.874396329s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:32.838228619 +0000 UTC m=+170.997423668" watchObservedRunningTime="2025-11-24 17:03:32.874396329 +0000 UTC m=+171.033591378" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.912945 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:32 crc kubenswrapper[4777]: E1124 17:03:32.916581 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:33.416568498 +0000 UTC m=+171.575763547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.947322 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-48b9w" podStartSLOduration=145.947303647 podStartE2EDuration="2m25.947303647s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:32.91629735 +0000 UTC m=+171.075492399" watchObservedRunningTime="2025-11-24 17:03:32.947303647 +0000 UTC m=+171.106498696" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.995845 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-9cm8z" podStartSLOduration=145.995824213 podStartE2EDuration="2m25.995824213s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:32.994955368 +0000 UTC m=+171.154150427" watchObservedRunningTime="2025-11-24 17:03:32.995824213 +0000 UTC m=+171.155019262" Nov 24 17:03:32 crc kubenswrapper[4777]: I1124 17:03:32.997899 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" podStartSLOduration=145.99788834 podStartE2EDuration="2m25.99788834s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:32.950245719 +0000 UTC m=+171.109440768" watchObservedRunningTime="2025-11-24 17:03:32.99788834 +0000 UTC m=+171.157083389" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.013916 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:33 crc kubenswrapper[4777]: E1124 17:03:33.014352 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:33.514322629 +0000 UTC m=+171.673517678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.038990 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" podStartSLOduration=146.038962478 podStartE2EDuration="2m26.038962478s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:33.038189606 +0000 UTC m=+171.197384655" watchObservedRunningTime="2025-11-24 17:03:33.038962478 +0000 UTC m=+171.198157527" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.115587 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:33 crc kubenswrapper[4777]: E1124 17:03:33.116029 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:33.61599679 +0000 UTC m=+171.775191839 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.144217 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.149616 4777 patch_prober.go:28] interesting pod/router-default-5444994796-9cm8z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:03:33 crc kubenswrapper[4777]: [-]has-synced failed: reason withheld Nov 24 17:03:33 crc kubenswrapper[4777]: [+]process-running ok Nov 24 17:03:33 crc kubenswrapper[4777]: healthz check failed Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.149685 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9cm8z" podUID="eacac718-a976-4e93-a7af-0414844177fb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.152723 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-bcj2h" podStartSLOduration=5.152711586 podStartE2EDuration="5.152711586s" podCreationTimestamp="2025-11-24 17:03:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:33.119345884 +0000 UTC m=+171.278540953" watchObservedRunningTime="2025-11-24 17:03:33.152711586 +0000 UTC m=+171.311906635" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.196957 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-qp69m" podStartSLOduration=146.196919492 podStartE2EDuration="2m26.196919492s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:33.154291981 +0000 UTC m=+171.313487040" watchObservedRunningTime="2025-11-24 17:03:33.196919492 +0000 UTC m=+171.356114561" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.219167 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:33 crc kubenswrapper[4777]: E1124 17:03:33.219598 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:33.719579845 +0000 UTC m=+171.878774894 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.302186 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-vk8pz" podStartSLOduration=146.302170863 podStartE2EDuration="2m26.302170863s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:33.196375716 +0000 UTC m=+171.355570765" watchObservedRunningTime="2025-11-24 17:03:33.302170863 +0000 UTC m=+171.461365912" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.324155 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:33 crc kubenswrapper[4777]: E1124 17:03:33.325430 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:33.825412402 +0000 UTC m=+171.984607451 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.426102 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:33 crc kubenswrapper[4777]: E1124 17:03:33.426554 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:33.926537788 +0000 UTC m=+172.085732837 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.484568 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.528054 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:33 crc kubenswrapper[4777]: E1124 17:03:33.528316 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:34.028304951 +0000 UTC m=+172.187500000 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.629697 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:33 crc kubenswrapper[4777]: E1124 17:03:33.630631 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:34.13059192 +0000 UTC m=+172.289786979 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.732097 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:33 crc kubenswrapper[4777]: E1124 17:03:33.732409 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:34.232393854 +0000 UTC m=+172.391588903 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.835334 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:33 crc kubenswrapper[4777]: E1124 17:03:33.836010 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:34.335993509 +0000 UTC m=+172.495188558 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.874947 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" event={"ID":"73fe25f4-c8e5-4d1e-a26e-25a3f0689593","Type":"ContainerStarted","Data":"75a110bf5f45338e9c52e7ff0a20409d4155a51249adb9bbbfdfb8cc5eb46888"} Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.875033 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.876903 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wbx6n" event={"ID":"194ce833-4172-4abf-929c-2a9c296a9675","Type":"ContainerStarted","Data":"296ef14fd03a705a4f6ec5d7d998937f223c38868e15d5d016fd6f714aab70ac"} Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.878243 4777 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-dtp89 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.878287 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" podUID="73fe25f4-c8e5-4d1e-a26e-25a3f0689593" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.881154 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l" event={"ID":"86b3f820-9431-4cee-bcac-bb17c9a43fab","Type":"ContainerStarted","Data":"93205007490be811c689ffc2ae959e9bed85633d9ecd3c0c7360c2ab5fce63ce"} Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.889697 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-chvs8" event={"ID":"404e867b-354a-45f7-b27a-10a7f3842555","Type":"ContainerStarted","Data":"1f094a02d720c429d22cb1c1f14deb8399301b48f2e894b014b7b2bb4f765942"} Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.902345 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" event={"ID":"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab","Type":"ContainerStarted","Data":"e6d9e2fde864cd94bc88afcc48e3729605ae1e54de574f289c93c7dff087f737"} Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.902390 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" event={"ID":"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab","Type":"ContainerStarted","Data":"486b6adbef148a2d440eb69531d54e8c3cbfb18fb236123ccf99497bd5a95b33"} Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.903928 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.908310 4777 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-6tcwx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" start-of-body= Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.908378 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" podUID="e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.909046 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" event={"ID":"4c5288e3-5eeb-4107-8222-cd847fe099d5","Type":"ContainerStarted","Data":"5029d27e102111e2549c2556cc393378661120e4c50876d22dc598555c66f485"} Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.909794 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.912639 4777 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-78g8k container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.912689 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" podUID="4c5288e3-5eeb-4107-8222-cd847fe099d5" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.916792 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" podStartSLOduration=146.916770265 podStartE2EDuration="2m26.916770265s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:33.908774362 +0000 UTC m=+172.067969411" watchObservedRunningTime="2025-11-24 17:03:33.916770265 +0000 UTC m=+172.075965314" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.927064 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zmdsl" event={"ID":"35059f02-39e8-40b0-8295-8230591cd996","Type":"ContainerStarted","Data":"c576e34d4ea8fd75e21ae5e2142f78685852eec91a48b1de482f07bb39a5fdf7"} Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.937664 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:33 crc kubenswrapper[4777]: E1124 17:03:33.940379 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:34.440366785 +0000 UTC m=+172.599561834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.949382 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb" event={"ID":"e0b913a5-16c4-4b8e-aa4e-0e04d9928537","Type":"ContainerStarted","Data":"352887adac90af39da09f40ba10f91bfad13206a6c5b5ff7e95b196640e92675"} Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.949427 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb" event={"ID":"e0b913a5-16c4-4b8e-aa4e-0e04d9928537","Type":"ContainerStarted","Data":"ffe7ee429f001c52b076d96a174eac0e4dc7035fc7eb84a6189c77e77a6949b7"} Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.950059 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.950840 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9bj8l" podStartSLOduration=146.950829727 podStartE2EDuration="2m26.950829727s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:33.93090752 +0000 UTC m=+172.090102569" watchObservedRunningTime="2025-11-24 17:03:33.950829727 +0000 UTC m=+172.110024776" Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.966299 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-lzzxl" event={"ID":"ac194a53-ef69-4048-bc97-200861512652","Type":"ContainerStarted","Data":"1bef0f332536fbdc491386ec4cfb1f31266e913f45eafcec0dc865523bf659cf"} Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.966377 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-lzzxl" event={"ID":"ac194a53-ef69-4048-bc97-200861512652","Type":"ContainerStarted","Data":"66ae89cd2672a924796d425d0489baf595020dfb8835e858e49fa4fa9176d739"} Nov 24 17:03:33 crc kubenswrapper[4777]: I1124 17:03:33.983161 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl" event={"ID":"9892a320-3e31-48df-bf8d-98441dd429ca","Type":"ContainerStarted","Data":"702a9ce5b2be647c08432f5bec8c644512dea820c6529d6eeee672d7331759dd"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.017913 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" podStartSLOduration=147.017888771 podStartE2EDuration="2m27.017888771s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:33.951814315 +0000 UTC m=+172.111009364" watchObservedRunningTime="2025-11-24 17:03:34.017888771 +0000 UTC m=+172.177083830" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.021086 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" event={"ID":"8062b545-bbdd-4a99-97c6-3682da1cb40f","Type":"ContainerStarted","Data":"7e58b00439f431eb17fc9362418c56937ee4167693bd267c1abafaa78ba00351"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.042050 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:34 crc kubenswrapper[4777]: E1124 17:03:34.044532 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:34.544479354 +0000 UTC m=+172.703674403 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.053894 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" podStartSLOduration=147.053876806 podStartE2EDuration="2m27.053876806s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.015786662 +0000 UTC m=+172.174981711" watchObservedRunningTime="2025-11-24 17:03:34.053876806 +0000 UTC m=+172.213071845" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.063231 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" event={"ID":"960e1bed-abef-4f45-9475-37eee6bd780a","Type":"ContainerStarted","Data":"850fac8ee9f39f5437be29b55178b02dfa83838a8cde4de4453e8385972b819a"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.063283 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kxqzn" event={"ID":"960e1bed-abef-4f45-9475-37eee6bd780a","Type":"ContainerStarted","Data":"69d7240c62d552b51d24e53b1e0b9610a91bf9e8636ead729a20b139f321af88"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.089677 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zmdsl" podStartSLOduration=147.089654986 podStartE2EDuration="2m27.089654986s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.08833983 +0000 UTC m=+172.247534879" watchObservedRunningTime="2025-11-24 17:03:34.089654986 +0000 UTC m=+172.248850035" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.091115 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb" podStartSLOduration=147.091105927 podStartE2EDuration="2m27.091105927s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.055246665 +0000 UTC m=+172.214441714" watchObservedRunningTime="2025-11-24 17:03:34.091105927 +0000 UTC m=+172.250300976" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.117445 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx" event={"ID":"f56db4a4-25ec-4ebe-b290-e46d122a556c","Type":"ContainerStarted","Data":"eea3d134a40eff0e56f69f95aa520c0b6f93f2c34776841b28077f8d991693b4"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.119324 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" event={"ID":"ff9e4751-a4e2-43c7-b75f-29ea3f7beca9","Type":"ContainerStarted","Data":"f793f71f0ccc4c2882b8359dd09ad263fa68170d71fb8d93b19b3088801f401f"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.124400 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" event={"ID":"55feb0bc-e9c4-4142-ac5d-a5e627bfd2d4","Type":"ContainerStarted","Data":"8b0414dc7983e68b58a7d3b9ba79bcaeb4b7de902ceebbcc2d85c4d0ac3f2360"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.140821 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" event={"ID":"7510a625-89f5-43c1-9ea2-7b836c01c676","Type":"ContainerStarted","Data":"dcab5e57ef173126bd91a3c6d1254115b2a01c8c88e776a6d2e1428c7e891d6b"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.140886 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" event={"ID":"7510a625-89f5-43c1-9ea2-7b836c01c676","Type":"ContainerStarted","Data":"ba5e9a25ba5e6fd43ebc1cd83b026f58cc0e591af91e617fd768f0d0d52aa10c"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.145242 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:34 crc kubenswrapper[4777]: E1124 17:03:34.145855 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:34.645831666 +0000 UTC m=+172.805026715 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.146430 4777 patch_prober.go:28] interesting pod/router-default-5444994796-9cm8z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:03:34 crc kubenswrapper[4777]: [-]has-synced failed: reason withheld Nov 24 17:03:34 crc kubenswrapper[4777]: [+]process-running ok Nov 24 17:03:34 crc kubenswrapper[4777]: healthz check failed Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.146485 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9cm8z" podUID="eacac718-a976-4e93-a7af-0414844177fb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.152028 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncqcl" podStartSLOduration=147.152010619 podStartE2EDuration="2m27.152010619s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.15170367 +0000 UTC m=+172.310898729" watchObservedRunningTime="2025-11-24 17:03:34.152010619 +0000 UTC m=+172.311205668" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.168843 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9qdq5" event={"ID":"3369a23a-c296-4fda-8318-85dd8cea9913","Type":"ContainerStarted","Data":"f39128356eed0ccd687259525ae6c6b87985cae247d1437ac93afc1a9808cca7"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.169313 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9qdq5" event={"ID":"3369a23a-c296-4fda-8318-85dd8cea9913","Type":"ContainerStarted","Data":"c88a8f972b9bb20ba884a20bb976bcd823d8294705b62cb2bc3fccb2007201d7"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.192348 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq" event={"ID":"493ed584-ded6-4027-ab8d-3946ecc089a4","Type":"ContainerStarted","Data":"8c5573b96200a99161ef19beb46e824ff8b3906df24d77636a1c666506813ba0"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.208202 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" podStartSLOduration=147.208188398 podStartE2EDuration="2m27.208188398s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.206217403 +0000 UTC m=+172.365412452" watchObservedRunningTime="2025-11-24 17:03:34.208188398 +0000 UTC m=+172.367383447" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.211104 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" event={"ID":"0f99ce69-4367-49eb-84c8-534959fa1735","Type":"ContainerStarted","Data":"1bb5f7d55f7527467196221c8631ad9d08f7bfa2ea286dc44c1ac3aa42c55f15"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.217404 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6" event={"ID":"903c1f59-fc19-48a1-941e-d67c125e0d42","Type":"ContainerStarted","Data":"d8de6263bb14eed7b82aeb8a459256885dfc429da90897c61425f23a9f813796"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.228445 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" event={"ID":"af533a06-13ef-4813-979d-875ffa3639dc","Type":"ContainerStarted","Data":"0d1d0bc5aa4c129ba64367e5b5c2a0776108d898370cb461465d483e567804c7"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.242117 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2" event={"ID":"2fd0dd98-6b46-4717-80c3-32d9502cae19","Type":"ContainerStarted","Data":"4949f62dbd7af2961cabad2d6b98290a1e156014d104ed09c8718e86019574f6"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.242158 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2" event={"ID":"2fd0dd98-6b46-4717-80c3-32d9502cae19","Type":"ContainerStarted","Data":"2709c7f5f1a64390a6bf06e53ef432125b37ad87f264dfa0d3050fbb6944d53b"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.248829 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:34 crc kubenswrapper[4777]: E1124 17:03:34.250266 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:34.750246304 +0000 UTC m=+172.909441353 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.252325 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" event={"ID":"54a60a6f-27d1-44d8-8372-512191c72f89","Type":"ContainerStarted","Data":"234cb8adac9b59e0d6c2b6d4eebda3dd87c18a72f2a002643a0dc3e8742eae0b"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.253414 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.259697 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" event={"ID":"7ec3567f-1218-441b-b33d-054f0c6a2e2d","Type":"ContainerStarted","Data":"ff812fb720a45f50229f29dee56fb02172ef403278db9ed610613eae9089da73"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.260729 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-lzzxl" podStartSLOduration=6.260712216 podStartE2EDuration="6.260712216s" podCreationTimestamp="2025-11-24 17:03:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.229886255 +0000 UTC m=+172.389081304" watchObservedRunningTime="2025-11-24 17:03:34.260712216 +0000 UTC m=+172.419907255" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.260912 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-kxqzn" podStartSLOduration=147.260904931 podStartE2EDuration="2m27.260904931s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.259346708 +0000 UTC m=+172.418541767" watchObservedRunningTime="2025-11-24 17:03:34.260904931 +0000 UTC m=+172.420099990" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.269701 4777 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-7hwq6 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" start-of-body= Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.269757 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" podUID="54a60a6f-27d1-44d8-8372-512191c72f89" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.270173 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" event={"ID":"9773b6b0-ffe7-40ed-b714-c7266b46e298","Type":"ContainerStarted","Data":"da2f196fb92c46abfe2c2ef899badcea0b32cdb68db4cb7964e93ef3f06d6dca"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.271175 4777 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-kd6cl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.271200 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" podUID="9773b6b0-ffe7-40ed-b714-c7266b46e298" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.274539 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" event={"ID":"eb260e29-56b9-4d87-9cee-7a8a3c336282","Type":"ContainerStarted","Data":"c9686e4705b3ccc28e89b3a55cd4f43455e35d6919f161f2ba71bf4037b5b282"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.275135 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.286013 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" podStartSLOduration=147.285994833 podStartE2EDuration="2m27.285994833s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.283446251 +0000 UTC m=+172.442641300" watchObservedRunningTime="2025-11-24 17:03:34.285994833 +0000 UTC m=+172.445189892" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.295490 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6b6gc" event={"ID":"0c79743f-3220-44ba-bc82-72ca21ee0daa","Type":"ContainerStarted","Data":"7db67d652fab1446791635371652778864b8f8b5335d01acd972418dc3a7efe8"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.302074 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv" event={"ID":"4f140bda-5ace-46c7-a5a7-242b6bf58207","Type":"ContainerStarted","Data":"dacad451af9e06f905a4d24a3506b77b7616cebd87b2296ee05254d7b0861d42"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.307671 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-f6tp2" podStartSLOduration=147.307649488 podStartE2EDuration="2m27.307649488s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.306045793 +0000 UTC m=+172.465240832" watchObservedRunningTime="2025-11-24 17:03:34.307649488 +0000 UTC m=+172.466844537" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.318523 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-82dhj" event={"ID":"2ada3f56-91fc-4155-a12a-dfe0d898d282","Type":"ContainerStarted","Data":"6281ed97cd462f045dc4b1606752bce8b5b4ff777528151791ca7a5de6c61515"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.318559 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-82dhj" event={"ID":"2ada3f56-91fc-4155-a12a-dfe0d898d282","Type":"ContainerStarted","Data":"4539575d6ab1c8ea177767abca3c6bf98f542e8be910e745cf158b23be0af0b7"} Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.319896 4777 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk8pz container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.319936 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vk8pz" podUID="90c90945-24a7-4f17-871c-7358792f0db5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.335760 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6dsxq" podStartSLOduration=147.335737122 podStartE2EDuration="2m27.335737122s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.335553027 +0000 UTC m=+172.494748076" watchObservedRunningTime="2025-11-24 17:03:34.335737122 +0000 UTC m=+172.494932161" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.353030 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:34 crc kubenswrapper[4777]: E1124 17:03:34.358152 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:34.858134548 +0000 UTC m=+173.017329597 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.369735 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-d6bzx" podStartSLOduration=147.369717412 podStartE2EDuration="2m27.369717412s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.368436106 +0000 UTC m=+172.527631155" watchObservedRunningTime="2025-11-24 17:03:34.369717412 +0000 UTC m=+172.528912461" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.392922 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w57nx" podStartSLOduration=147.39290616 podStartE2EDuration="2m27.39290616s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.392151769 +0000 UTC m=+172.551346818" watchObservedRunningTime="2025-11-24 17:03:34.39290616 +0000 UTC m=+172.552101199" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.421570 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7dhx6" podStartSLOduration=147.42154318 podStartE2EDuration="2m27.42154318s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.42119568 +0000 UTC m=+172.580390729" watchObservedRunningTime="2025-11-24 17:03:34.42154318 +0000 UTC m=+172.580738229" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.457355 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-s86qg" podStartSLOduration=147.45733376 podStartE2EDuration="2m27.45733376s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.456929509 +0000 UTC m=+172.616124558" watchObservedRunningTime="2025-11-24 17:03:34.45733376 +0000 UTC m=+172.616528809" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.457644 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:34 crc kubenswrapper[4777]: E1124 17:03:34.457711 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:34.95769753 +0000 UTC m=+173.116892579 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.459722 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:34 crc kubenswrapper[4777]: E1124 17:03:34.460134 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:34.960122498 +0000 UTC m=+173.119317547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.513586 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" podStartSLOduration=147.513568052 podStartE2EDuration="2m27.513568052s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.494137309 +0000 UTC m=+172.653332358" watchObservedRunningTime="2025-11-24 17:03:34.513568052 +0000 UTC m=+172.672763101" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.563143 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:34 crc kubenswrapper[4777]: E1124 17:03:34.563651 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.063636401 +0000 UTC m=+173.222831450 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.568884 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x2k8g" podStartSLOduration=147.568865407 podStartE2EDuration="2m27.568865407s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.567811007 +0000 UTC m=+172.727006066" watchObservedRunningTime="2025-11-24 17:03:34.568865407 +0000 UTC m=+172.728060456" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.572434 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-8fgbp" podStartSLOduration=147.572418326 podStartE2EDuration="2m27.572418326s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.516469293 +0000 UTC m=+172.675664342" watchObservedRunningTime="2025-11-24 17:03:34.572418326 +0000 UTC m=+172.731613385" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.610473 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.611875 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-9qdq5" podStartSLOduration=147.611864908 podStartE2EDuration="2m27.611864908s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.610884981 +0000 UTC m=+172.770080050" watchObservedRunningTime="2025-11-24 17:03:34.611864908 +0000 UTC m=+172.771059957" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.665691 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:34 crc kubenswrapper[4777]: E1124 17:03:34.666060 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.166044132 +0000 UTC m=+173.325239181 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.685787 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6b6gc" podStartSLOduration=147.685756243 podStartE2EDuration="2m27.685756243s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.641357622 +0000 UTC m=+172.800552681" watchObservedRunningTime="2025-11-24 17:03:34.685756243 +0000 UTC m=+172.844951292" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.686319 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" podStartSLOduration=147.686310589 podStartE2EDuration="2m27.686310589s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.683396397 +0000 UTC m=+172.842591446" watchObservedRunningTime="2025-11-24 17:03:34.686310589 +0000 UTC m=+172.845505638" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.764512 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" podStartSLOduration=147.764494043 podStartE2EDuration="2m27.764494043s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.762450306 +0000 UTC m=+172.921645345" watchObservedRunningTime="2025-11-24 17:03:34.764494043 +0000 UTC m=+172.923689092" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.768103 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:34 crc kubenswrapper[4777]: E1124 17:03:34.768450 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.268436003 +0000 UTC m=+173.427631052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.794603 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5cdnv" podStartSLOduration=147.794573284 podStartE2EDuration="2m27.794573284s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.792092144 +0000 UTC m=+172.951287193" watchObservedRunningTime="2025-11-24 17:03:34.794573284 +0000 UTC m=+172.953768333" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.835064 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-82dhj" podStartSLOduration=147.835032984 podStartE2EDuration="2m27.835032984s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:34.83236888 +0000 UTC m=+172.991563929" watchObservedRunningTime="2025-11-24 17:03:34.835032984 +0000 UTC m=+172.994228033" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.849800 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.849855 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.871127 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:34 crc kubenswrapper[4777]: E1124 17:03:34.871727 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.371712489 +0000 UTC m=+173.530907538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.972474 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:34 crc kubenswrapper[4777]: E1124 17:03:34.972698 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.47266807 +0000 UTC m=+173.631863119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:34 crc kubenswrapper[4777]: I1124 17:03:34.972983 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:34 crc kubenswrapper[4777]: E1124 17:03:34.973323 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.473315908 +0000 UTC m=+173.632510957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.075382 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.075541 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.575516874 +0000 UTC m=+173.734711923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.075754 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.076097 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.57608971 +0000 UTC m=+173.735284759 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.084375 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rrgd4" Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.084435 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.084466 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.086745 4777 patch_prober.go:28] interesting pod/apiserver-76f77b778f-qz2kt container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.8:8443/livez\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.086821 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" podUID="8062b545-bbdd-4a99-97c6-3682da1cb40f" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.8:8443/livez\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.149197 4777 patch_prober.go:28] interesting pod/router-default-5444994796-9cm8z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:03:35 crc kubenswrapper[4777]: [-]has-synced failed: reason withheld Nov 24 17:03:35 crc kubenswrapper[4777]: [+]process-running ok Nov 24 17:03:35 crc kubenswrapper[4777]: healthz check failed Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.149254 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9cm8z" podUID="eacac718-a976-4e93-a7af-0414844177fb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.176561 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.176748 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.676722322 +0000 UTC m=+173.835917371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.176885 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.177212 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.677200505 +0000 UTC m=+173.836395554 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.278412 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.278518 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.778503726 +0000 UTC m=+173.937698775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.278752 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.279127 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.779119983 +0000 UTC m=+173.938315032 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.335056 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-chvs8" event={"ID":"404e867b-354a-45f7-b27a-10a7f3842555","Type":"ContainerStarted","Data":"d6461dba7889092b78930cb8d1a7cceb85087b5aa7735e58ad21d8a498a8e7cf"} Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.337181 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wbx6n" event={"ID":"194ce833-4172-4abf-929c-2a9c296a9675","Type":"ContainerStarted","Data":"7be2aa630e533bcf22361f72dda53e63aca434ba31943a9de645bb1e05169468"} Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.337347 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-wbx6n" Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.339586 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" event={"ID":"4ef71d5e-7018-4d3b-9eb1-abba188e4a39","Type":"ContainerStarted","Data":"b47213aaba73907215fe5a4c54fd360b7eb9d6743e735cc09edcc6b7c35d730a"} Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.351166 4777 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-6tcwx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" start-of-body= Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.351218 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" podUID="e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.358472 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.361074 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dtp89" Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.383343 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.383621 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.883557172 +0000 UTC m=+174.042752221 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.388312 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.390206 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.890194297 +0000 UTC m=+174.049389346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.442705 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-78g8k" Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.462834 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-chvs8" podStartSLOduration=148.462816436 podStartE2EDuration="2m28.462816436s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:35.382471681 +0000 UTC m=+173.541666730" watchObservedRunningTime="2025-11-24 17:03:35.462816436 +0000 UTC m=+173.622011485" Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.489659 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.489820 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.98980089 +0000 UTC m=+174.148995939 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.489956 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.491332 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:35.991307942 +0000 UTC m=+174.150502991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.545910 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-wbx6n" podStartSLOduration=8.545877567 podStartE2EDuration="8.545877567s" podCreationTimestamp="2025-11-24 17:03:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:35.541777763 +0000 UTC m=+173.700972812" watchObservedRunningTime="2025-11-24 17:03:35.545877567 +0000 UTC m=+173.705072606" Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.591128 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.591413 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:36.091376179 +0000 UTC m=+174.250571368 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.694109 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.694499 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:36.19448781 +0000 UTC m=+174.353682859 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.795593 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.795710 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:36.295691837 +0000 UTC m=+174.454886886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.796002 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.796290 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:36.296282124 +0000 UTC m=+174.455477173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.847176 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.896810 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.896926 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:36.396911126 +0000 UTC m=+174.556106175 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.897286 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:35 crc kubenswrapper[4777]: E1124 17:03:35.897560 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:36.397553204 +0000 UTC m=+174.556748253 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.911110 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:03:35 crc kubenswrapper[4777]: I1124 17:03:35.999199 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:36 crc kubenswrapper[4777]: E1124 17:03:35.999606 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:36.499591445 +0000 UTC m=+174.658786494 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.101045 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:36 crc kubenswrapper[4777]: E1124 17:03:36.101360 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:36.601349669 +0000 UTC m=+174.760544718 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.151158 4777 patch_prober.go:28] interesting pod/router-default-5444994796-9cm8z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:03:36 crc kubenswrapper[4777]: [-]has-synced failed: reason withheld Nov 24 17:03:36 crc kubenswrapper[4777]: [+]process-running ok Nov 24 17:03:36 crc kubenswrapper[4777]: healthz check failed Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.151210 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9cm8z" podUID="eacac718-a976-4e93-a7af-0414844177fb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.202144 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:36 crc kubenswrapper[4777]: E1124 17:03:36.202641 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:36.702626688 +0000 UTC m=+174.861821737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.304292 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:36 crc kubenswrapper[4777]: E1124 17:03:36.304636 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:36.804625168 +0000 UTC m=+174.963820217 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.345559 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" event={"ID":"4ef71d5e-7018-4d3b-9eb1-abba188e4a39","Type":"ContainerStarted","Data":"2e800f6c077b064142c709466fa50279af526b3f26ba2a78d3d8db5ae0e2599a"} Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.348533 4777 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-6tcwx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" start-of-body= Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.348586 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" podUID="e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.352372 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlq9c" Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.405381 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:36 crc kubenswrapper[4777]: E1124 17:03:36.405558 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:36.905533738 +0000 UTC m=+175.064728787 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.405814 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:36 crc kubenswrapper[4777]: E1124 17:03:36.406899 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:36.906885076 +0000 UTC m=+175.066080135 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.507909 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:36 crc kubenswrapper[4777]: E1124 17:03:36.508345 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:37.0083291 +0000 UTC m=+175.167524149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.539426 4777 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.610046 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:36 crc kubenswrapper[4777]: E1124 17:03:36.610723 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:37.110689281 +0000 UTC m=+175.269884520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.711496 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:36 crc kubenswrapper[4777]: E1124 17:03:36.712081 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:37.212058323 +0000 UTC m=+175.371253372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.813131 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:36 crc kubenswrapper[4777]: E1124 17:03:36.813413 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:37.313400115 +0000 UTC m=+175.472595164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.813996 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-shz64"] Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.814940 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.816827 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.843507 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-shz64"] Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.915480 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.915664 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8ng5\" (UniqueName: \"kubernetes.io/projected/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-kube-api-access-w8ng5\") pod \"certified-operators-shz64\" (UID: \"b93bcc12-1487-4c0f-9da2-fab9c61df5c9\") " pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.915698 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-utilities\") pod \"certified-operators-shz64\" (UID: \"b93bcc12-1487-4c0f-9da2-fab9c61df5c9\") " pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.915715 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-catalog-content\") pod \"certified-operators-shz64\" (UID: \"b93bcc12-1487-4c0f-9da2-fab9c61df5c9\") " pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:03:36 crc kubenswrapper[4777]: E1124 17:03:36.915838 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:37.415824327 +0000 UTC m=+175.575019376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.986353 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tr7gm"] Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.987352 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:03:36 crc kubenswrapper[4777]: I1124 17:03:36.989845 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.017778 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8ng5\" (UniqueName: \"kubernetes.io/projected/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-kube-api-access-w8ng5\") pod \"certified-operators-shz64\" (UID: \"b93bcc12-1487-4c0f-9da2-fab9c61df5c9\") " pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.017825 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-utilities\") pod \"certified-operators-shz64\" (UID: \"b93bcc12-1487-4c0f-9da2-fab9c61df5c9\") " pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.017846 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-catalog-content\") pod \"certified-operators-shz64\" (UID: \"b93bcc12-1487-4c0f-9da2-fab9c61df5c9\") " pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.017898 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:37 crc kubenswrapper[4777]: E1124 17:03:37.018208 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:37.518195288 +0000 UTC m=+175.677390327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.018702 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-utilities\") pod \"certified-operators-shz64\" (UID: \"b93bcc12-1487-4c0f-9da2-fab9c61df5c9\") " pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.018733 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-catalog-content\") pod \"certified-operators-shz64\" (UID: \"b93bcc12-1487-4c0f-9da2-fab9c61df5c9\") " pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.055556 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8ng5\" (UniqueName: \"kubernetes.io/projected/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-kube-api-access-w8ng5\") pod \"certified-operators-shz64\" (UID: \"b93bcc12-1487-4c0f-9da2-fab9c61df5c9\") " pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.092418 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tr7gm"] Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.109741 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cw68j" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.118669 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:37 crc kubenswrapper[4777]: E1124 17:03:37.118836 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:37.618813299 +0000 UTC m=+175.778008348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.118962 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6203d413-faf0-4476-9a91-345ba4f1772f-catalog-content\") pod \"community-operators-tr7gm\" (UID: \"6203d413-faf0-4476-9a91-345ba4f1772f\") " pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.119051 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6203d413-faf0-4476-9a91-345ba4f1772f-utilities\") pod \"community-operators-tr7gm\" (UID: \"6203d413-faf0-4476-9a91-345ba4f1772f\") " pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.119103 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6qnd\" (UniqueName: \"kubernetes.io/projected/6203d413-faf0-4476-9a91-345ba4f1772f-kube-api-access-b6qnd\") pod \"community-operators-tr7gm\" (UID: \"6203d413-faf0-4476-9a91-345ba4f1772f\") " pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.119256 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:37 crc kubenswrapper[4777]: E1124 17:03:37.119584 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:37.619571801 +0000 UTC m=+175.778766850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.135128 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.152816 4777 patch_prober.go:28] interesting pod/router-default-5444994796-9cm8z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:03:37 crc kubenswrapper[4777]: [-]has-synced failed: reason withheld Nov 24 17:03:37 crc kubenswrapper[4777]: [+]process-running ok Nov 24 17:03:37 crc kubenswrapper[4777]: healthz check failed Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.152879 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9cm8z" podUID="eacac718-a976-4e93-a7af-0414844177fb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.189305 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wmfnz"] Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.192509 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.222820 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.225602 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wmfnz"] Nov 24 17:03:37 crc kubenswrapper[4777]: E1124 17:03:37.230937 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:37.730906292 +0000 UTC m=+175.890101331 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.232436 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6203d413-faf0-4476-9a91-345ba4f1772f-catalog-content\") pod \"community-operators-tr7gm\" (UID: \"6203d413-faf0-4476-9a91-345ba4f1772f\") " pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.232517 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6203d413-faf0-4476-9a91-345ba4f1772f-utilities\") pod \"community-operators-tr7gm\" (UID: \"6203d413-faf0-4476-9a91-345ba4f1772f\") " pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.232593 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6qnd\" (UniqueName: \"kubernetes.io/projected/6203d413-faf0-4476-9a91-345ba4f1772f-kube-api-access-b6qnd\") pod \"community-operators-tr7gm\" (UID: \"6203d413-faf0-4476-9a91-345ba4f1772f\") " pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.232656 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:37 crc kubenswrapper[4777]: E1124 17:03:37.233016 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:37.73300322 +0000 UTC m=+175.892198269 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.233148 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6203d413-faf0-4476-9a91-345ba4f1772f-utilities\") pod \"community-operators-tr7gm\" (UID: \"6203d413-faf0-4476-9a91-345ba4f1772f\") " pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.233477 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6203d413-faf0-4476-9a91-345ba4f1772f-catalog-content\") pod \"community-operators-tr7gm\" (UID: \"6203d413-faf0-4476-9a91-345ba4f1772f\") " pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.279990 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6qnd\" (UniqueName: \"kubernetes.io/projected/6203d413-faf0-4476-9a91-345ba4f1772f-kube-api-access-b6qnd\") pod \"community-operators-tr7gm\" (UID: \"6203d413-faf0-4476-9a91-345ba4f1772f\") " pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.302529 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.334208 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.334609 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-utilities\") pod \"certified-operators-wmfnz\" (UID: \"eebaa807-fbd5-40d7-ac22-02523cbc4fc2\") " pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.334877 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpdcz\" (UniqueName: \"kubernetes.io/projected/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-kube-api-access-fpdcz\") pod \"certified-operators-wmfnz\" (UID: \"eebaa807-fbd5-40d7-ac22-02523cbc4fc2\") " pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.335024 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-catalog-content\") pod \"certified-operators-wmfnz\" (UID: \"eebaa807-fbd5-40d7-ac22-02523cbc4fc2\") " pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:03:37 crc kubenswrapper[4777]: E1124 17:03:37.335201 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 17:03:37.835174835 +0000 UTC m=+175.994369884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.356258 4777 generic.go:334] "Generic (PLEG): container finished" podID="7ec3567f-1218-441b-b33d-054f0c6a2e2d" containerID="ff812fb720a45f50229f29dee56fb02172ef403278db9ed610613eae9089da73" exitCode=0 Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.356327 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" event={"ID":"7ec3567f-1218-441b-b33d-054f0c6a2e2d","Type":"ContainerDied","Data":"ff812fb720a45f50229f29dee56fb02172ef403278db9ed610613eae9089da73"} Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.369173 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hqk2w"] Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.370549 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.392935 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hqk2w"] Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.395261 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" event={"ID":"4ef71d5e-7018-4d3b-9eb1-abba188e4a39","Type":"ContainerStarted","Data":"4f14281edb728a08260c7ecc92380994d0ea3248c8b1209ddd6e3bcebfa7c2a7"} Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.395298 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" event={"ID":"4ef71d5e-7018-4d3b-9eb1-abba188e4a39","Type":"ContainerStarted","Data":"48ccb7115cecdb1f9c8068853a749b14ebd6b56b52331149451d54a38a04046d"} Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.437989 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.438032 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-utilities\") pod \"certified-operators-wmfnz\" (UID: \"eebaa807-fbd5-40d7-ac22-02523cbc4fc2\") " pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.438048 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpdcz\" (UniqueName: \"kubernetes.io/projected/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-kube-api-access-fpdcz\") pod \"certified-operators-wmfnz\" (UID: \"eebaa807-fbd5-40d7-ac22-02523cbc4fc2\") " pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.438127 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48ca8fd0-3e8b-4df1-84c6-01104522f616-catalog-content\") pod \"community-operators-hqk2w\" (UID: \"48ca8fd0-3e8b-4df1-84c6-01104522f616\") " pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.438286 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-catalog-content\") pod \"certified-operators-wmfnz\" (UID: \"eebaa807-fbd5-40d7-ac22-02523cbc4fc2\") " pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.438385 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg67t\" (UniqueName: \"kubernetes.io/projected/48ca8fd0-3e8b-4df1-84c6-01104522f616-kube-api-access-gg67t\") pod \"community-operators-hqk2w\" (UID: \"48ca8fd0-3e8b-4df1-84c6-01104522f616\") " pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.438409 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48ca8fd0-3e8b-4df1-84c6-01104522f616-utilities\") pod \"community-operators-hqk2w\" (UID: \"48ca8fd0-3e8b-4df1-84c6-01104522f616\") " pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:03:37 crc kubenswrapper[4777]: E1124 17:03:37.440755 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 17:03:37.940740565 +0000 UTC m=+176.099935614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wsl2k" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.441462 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-utilities\") pod \"certified-operators-wmfnz\" (UID: \"eebaa807-fbd5-40d7-ac22-02523cbc4fc2\") " pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.446748 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-catalog-content\") pod \"certified-operators-wmfnz\" (UID: \"eebaa807-fbd5-40d7-ac22-02523cbc4fc2\") " pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.477224 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpdcz\" (UniqueName: \"kubernetes.io/projected/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-kube-api-access-fpdcz\") pod \"certified-operators-wmfnz\" (UID: \"eebaa807-fbd5-40d7-ac22-02523cbc4fc2\") " pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.490131 4777 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-24T17:03:36.53945929Z","Handler":null,"Name":""} Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.505200 4777 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.505242 4777 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.523768 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.541693 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.541914 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg67t\" (UniqueName: \"kubernetes.io/projected/48ca8fd0-3e8b-4df1-84c6-01104522f616-kube-api-access-gg67t\") pod \"community-operators-hqk2w\" (UID: \"48ca8fd0-3e8b-4df1-84c6-01104522f616\") " pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.541936 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48ca8fd0-3e8b-4df1-84c6-01104522f616-utilities\") pod \"community-operators-hqk2w\" (UID: \"48ca8fd0-3e8b-4df1-84c6-01104522f616\") " pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.542559 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48ca8fd0-3e8b-4df1-84c6-01104522f616-catalog-content\") pod \"community-operators-hqk2w\" (UID: \"48ca8fd0-3e8b-4df1-84c6-01104522f616\") " pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.546345 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48ca8fd0-3e8b-4df1-84c6-01104522f616-utilities\") pod \"community-operators-hqk2w\" (UID: \"48ca8fd0-3e8b-4df1-84c6-01104522f616\") " pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.546911 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48ca8fd0-3e8b-4df1-84c6-01104522f616-catalog-content\") pod \"community-operators-hqk2w\" (UID: \"48ca8fd0-3e8b-4df1-84c6-01104522f616\") " pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.554369 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.588847 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg67t\" (UniqueName: \"kubernetes.io/projected/48ca8fd0-3e8b-4df1-84c6-01104522f616-kube-api-access-gg67t\") pod \"community-operators-hqk2w\" (UID: \"48ca8fd0-3e8b-4df1-84c6-01104522f616\") " pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.612009 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-fsbnq" podStartSLOduration=10.611987939 podStartE2EDuration="10.611987939s" podCreationTimestamp="2025-11-24 17:03:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:37.453373257 +0000 UTC m=+175.612568306" watchObservedRunningTime="2025-11-24 17:03:37.611987939 +0000 UTC m=+175.771182988" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.614527 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-shz64"] Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.645107 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:37 crc kubenswrapper[4777]: W1124 17:03:37.648735 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb93bcc12_1487_4c0f_9da2_fab9c61df5c9.slice/crio-c45cfe50b77cf43e9df4b5a9c22a188d037546d16dc6c03ab73633547571c1fc WatchSource:0}: Error finding container c45cfe50b77cf43e9df4b5a9c22a188d037546d16dc6c03ab73633547571c1fc: Status 404 returned error can't find the container with id c45cfe50b77cf43e9df4b5a9c22a188d037546d16dc6c03ab73633547571c1fc Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.655472 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.655529 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.730628 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.781833 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wsl2k\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.870873 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wmfnz"] Nov 24 17:03:37 crc kubenswrapper[4777]: I1124 17:03:37.921932 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tr7gm"] Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.032777 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.146572 4777 patch_prober.go:28] interesting pod/router-default-5444994796-9cm8z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:03:38 crc kubenswrapper[4777]: [-]has-synced failed: reason withheld Nov 24 17:03:38 crc kubenswrapper[4777]: [+]process-running ok Nov 24 17:03:38 crc kubenswrapper[4777]: healthz check failed Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.147038 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9cm8z" podUID="eacac718-a976-4e93-a7af-0414844177fb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.270622 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hqk2w"] Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.332927 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wsl2k"] Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.415510 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmfnz" event={"ID":"eebaa807-fbd5-40d7-ac22-02523cbc4fc2","Type":"ContainerDied","Data":"f1698cab81e104818a14e4285bcfa695c045c6f501c3461def680fe9a145c478"} Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.415633 4777 generic.go:334] "Generic (PLEG): container finished" podID="eebaa807-fbd5-40d7-ac22-02523cbc4fc2" containerID="f1698cab81e104818a14e4285bcfa695c045c6f501c3461def680fe9a145c478" exitCode=0 Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.415737 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmfnz" event={"ID":"eebaa807-fbd5-40d7-ac22-02523cbc4fc2","Type":"ContainerStarted","Data":"ae4299f1cc5dbfc5f9a6757920eb49cf791c404b6ba976e8fa8507cb12492137"} Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.433022 4777 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.436907 4777 generic.go:334] "Generic (PLEG): container finished" podID="6203d413-faf0-4476-9a91-345ba4f1772f" containerID="3c97b6302bf5a6f62b361f7f5a5c20b7bcb13a1b40d6502b4c78a62332bfdbeb" exitCode=0 Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.437027 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr7gm" event={"ID":"6203d413-faf0-4476-9a91-345ba4f1772f","Type":"ContainerDied","Data":"3c97b6302bf5a6f62b361f7f5a5c20b7bcb13a1b40d6502b4c78a62332bfdbeb"} Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.437057 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr7gm" event={"ID":"6203d413-faf0-4476-9a91-345ba4f1772f","Type":"ContainerStarted","Data":"7ff90c859ffd8ba86b369c0ac90a167b40bbd5de37eb9f298fc766383d228b4b"} Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.439336 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqk2w" event={"ID":"48ca8fd0-3e8b-4df1-84c6-01104522f616","Type":"ContainerStarted","Data":"d707497fa3ac4be390c2151332b8750c0a6d9c873995ecb6844ee55f6b7a7d62"} Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.441658 4777 generic.go:334] "Generic (PLEG): container finished" podID="b93bcc12-1487-4c0f-9da2-fab9c61df5c9" containerID="2bf66f64adac1f0f975225dd4b6f5be282bdbe508c9dbb3987a9633417089d23" exitCode=0 Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.441707 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-shz64" event={"ID":"b93bcc12-1487-4c0f-9da2-fab9c61df5c9","Type":"ContainerDied","Data":"2bf66f64adac1f0f975225dd4b6f5be282bdbe508c9dbb3987a9633417089d23"} Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.441726 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-shz64" event={"ID":"b93bcc12-1487-4c0f-9da2-fab9c61df5c9","Type":"ContainerStarted","Data":"c45cfe50b77cf43e9df4b5a9c22a188d037546d16dc6c03ab73633547571c1fc"} Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.456148 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" event={"ID":"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5","Type":"ContainerStarted","Data":"0c74eba9d1bc898dc08ec8ab2e8b7219955c87e6511d9bcb08c8de56f946a4d4"} Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.699005 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.760837 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn9ck\" (UniqueName: \"kubernetes.io/projected/7ec3567f-1218-441b-b33d-054f0c6a2e2d-kube-api-access-qn9ck\") pod \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\" (UID: \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\") " Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.760919 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ec3567f-1218-441b-b33d-054f0c6a2e2d-config-volume\") pod \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\" (UID: \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\") " Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.761072 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7ec3567f-1218-441b-b33d-054f0c6a2e2d-secret-volume\") pod \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\" (UID: \"7ec3567f-1218-441b-b33d-054f0c6a2e2d\") " Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.762086 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ec3567f-1218-441b-b33d-054f0c6a2e2d-config-volume" (OuterVolumeSpecName: "config-volume") pod "7ec3567f-1218-441b-b33d-054f0c6a2e2d" (UID: "7ec3567f-1218-441b-b33d-054f0c6a2e2d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.764681 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ec3567f-1218-441b-b33d-054f0c6a2e2d-kube-api-access-qn9ck" (OuterVolumeSpecName: "kube-api-access-qn9ck") pod "7ec3567f-1218-441b-b33d-054f0c6a2e2d" (UID: "7ec3567f-1218-441b-b33d-054f0c6a2e2d"). InnerVolumeSpecName "kube-api-access-qn9ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.764759 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ec3567f-1218-441b-b33d-054f0c6a2e2d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7ec3567f-1218-441b-b33d-054f0c6a2e2d" (UID: "7ec3567f-1218-441b-b33d-054f0c6a2e2d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.862259 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn9ck\" (UniqueName: \"kubernetes.io/projected/7ec3567f-1218-441b-b33d-054f0c6a2e2d-kube-api-access-qn9ck\") on node \"crc\" DevicePath \"\"" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.862305 4777 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ec3567f-1218-441b-b33d-054f0c6a2e2d-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.862315 4777 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7ec3567f-1218-441b-b33d-054f0c6a2e2d-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.920408 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 17:03:38 crc kubenswrapper[4777]: E1124 17:03:38.920664 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ec3567f-1218-441b-b33d-054f0c6a2e2d" containerName="collect-profiles" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.920679 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ec3567f-1218-441b-b33d-054f0c6a2e2d" containerName="collect-profiles" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.920780 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ec3567f-1218-441b-b33d-054f0c6a2e2d" containerName="collect-profiles" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.921139 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.923416 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.923607 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.942441 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.971934 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0493374-1b14-4216-bf96-b81a619212e5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e0493374-1b14-4216-bf96-b81a619212e5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.972086 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0493374-1b14-4216-bf96-b81a619212e5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e0493374-1b14-4216-bf96-b81a619212e5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.983065 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9mbg7"] Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.985523 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.989609 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mbg7"] Nov 24 17:03:38 crc kubenswrapper[4777]: I1124 17:03:38.991134 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.073455 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0493374-1b14-4216-bf96-b81a619212e5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e0493374-1b14-4216-bf96-b81a619212e5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.073617 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0493374-1b14-4216-bf96-b81a619212e5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e0493374-1b14-4216-bf96-b81a619212e5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.073630 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caacdae7-8c80-4fcb-85de-78dcc1cecf59-utilities\") pod \"redhat-marketplace-9mbg7\" (UID: \"caacdae7-8c80-4fcb-85de-78dcc1cecf59\") " pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.073723 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caacdae7-8c80-4fcb-85de-78dcc1cecf59-catalog-content\") pod \"redhat-marketplace-9mbg7\" (UID: \"caacdae7-8c80-4fcb-85de-78dcc1cecf59\") " pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.073761 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbflk\" (UniqueName: \"kubernetes.io/projected/caacdae7-8c80-4fcb-85de-78dcc1cecf59-kube-api-access-gbflk\") pod \"redhat-marketplace-9mbg7\" (UID: \"caacdae7-8c80-4fcb-85de-78dcc1cecf59\") " pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.073834 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0493374-1b14-4216-bf96-b81a619212e5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e0493374-1b14-4216-bf96-b81a619212e5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.094696 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0493374-1b14-4216-bf96-b81a619212e5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e0493374-1b14-4216-bf96-b81a619212e5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.147280 4777 patch_prober.go:28] interesting pod/router-default-5444994796-9cm8z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:03:39 crc kubenswrapper[4777]: [-]has-synced failed: reason withheld Nov 24 17:03:39 crc kubenswrapper[4777]: [+]process-running ok Nov 24 17:03:39 crc kubenswrapper[4777]: healthz check failed Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.147405 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9cm8z" podUID="eacac718-a976-4e93-a7af-0414844177fb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.175187 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caacdae7-8c80-4fcb-85de-78dcc1cecf59-utilities\") pod \"redhat-marketplace-9mbg7\" (UID: \"caacdae7-8c80-4fcb-85de-78dcc1cecf59\") " pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.175278 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caacdae7-8c80-4fcb-85de-78dcc1cecf59-catalog-content\") pod \"redhat-marketplace-9mbg7\" (UID: \"caacdae7-8c80-4fcb-85de-78dcc1cecf59\") " pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.175335 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbflk\" (UniqueName: \"kubernetes.io/projected/caacdae7-8c80-4fcb-85de-78dcc1cecf59-kube-api-access-gbflk\") pod \"redhat-marketplace-9mbg7\" (UID: \"caacdae7-8c80-4fcb-85de-78dcc1cecf59\") " pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.175839 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caacdae7-8c80-4fcb-85de-78dcc1cecf59-catalog-content\") pod \"redhat-marketplace-9mbg7\" (UID: \"caacdae7-8c80-4fcb-85de-78dcc1cecf59\") " pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.176114 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caacdae7-8c80-4fcb-85de-78dcc1cecf59-utilities\") pod \"redhat-marketplace-9mbg7\" (UID: \"caacdae7-8c80-4fcb-85de-78dcc1cecf59\") " pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.197502 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbflk\" (UniqueName: \"kubernetes.io/projected/caacdae7-8c80-4fcb-85de-78dcc1cecf59-kube-api-access-gbflk\") pod \"redhat-marketplace-9mbg7\" (UID: \"caacdae7-8c80-4fcb-85de-78dcc1cecf59\") " pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.242364 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.253719 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.314799 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.367475 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rtc8b"] Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.368694 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.380857 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtc8b"] Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.464777 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 17:03:39 crc kubenswrapper[4777]: W1124 17:03:39.477300 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode0493374_1b14_4216_bf96_b81a619212e5.slice/crio-6f0369278739c14cf61eab2f74237a0b7ba3dc68df2276ffe3fe145f1d390fd4 WatchSource:0}: Error finding container 6f0369278739c14cf61eab2f74237a0b7ba3dc68df2276ffe3fe145f1d390fd4: Status 404 returned error can't find the container with id 6f0369278739c14cf61eab2f74237a0b7ba3dc68df2276ffe3fe145f1d390fd4 Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.479532 4777 generic.go:334] "Generic (PLEG): container finished" podID="48ca8fd0-3e8b-4df1-84c6-01104522f616" containerID="607f5c5d9f5aec968a92b2544908277da64bba3896e2f515b22fe14d896adfde" exitCode=0 Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.479614 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqk2w" event={"ID":"48ca8fd0-3e8b-4df1-84c6-01104522f616","Type":"ContainerDied","Data":"607f5c5d9f5aec968a92b2544908277da64bba3896e2f515b22fe14d896adfde"} Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.479757 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/549af019-cce7-46bc-a02b-41e1c4e09256-catalog-content\") pod \"redhat-marketplace-rtc8b\" (UID: \"549af019-cce7-46bc-a02b-41e1c4e09256\") " pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.479820 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/549af019-cce7-46bc-a02b-41e1c4e09256-utilities\") pod \"redhat-marketplace-rtc8b\" (UID: \"549af019-cce7-46bc-a02b-41e1c4e09256\") " pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.479878 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqpg9\" (UniqueName: \"kubernetes.io/projected/549af019-cce7-46bc-a02b-41e1c4e09256-kube-api-access-cqpg9\") pod \"redhat-marketplace-rtc8b\" (UID: \"549af019-cce7-46bc-a02b-41e1c4e09256\") " pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.485694 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.486278 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw" event={"ID":"7ec3567f-1218-441b-b33d-054f0c6a2e2d","Type":"ContainerDied","Data":"d48f19487aea555851dd6772c9a9157e6bd3f6e58ec6be8abc03d42c6a184797"} Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.486301 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d48f19487aea555851dd6772c9a9157e6bd3f6e58ec6be8abc03d42c6a184797" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.489899 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" event={"ID":"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5","Type":"ContainerStarted","Data":"2ef6bb31c9d4f22edf3828875d522daa9702614edd54f87315bd5488d00e9060"} Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.490482 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.517713 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" podStartSLOduration=152.51769762 podStartE2EDuration="2m32.51769762s" podCreationTimestamp="2025-11-24 17:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:39.516149067 +0000 UTC m=+177.675344126" watchObservedRunningTime="2025-11-24 17:03:39.51769762 +0000 UTC m=+177.676892669" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.554319 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mbg7"] Nov 24 17:03:39 crc kubenswrapper[4777]: W1124 17:03:39.569776 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcaacdae7_8c80_4fcb_85de_78dcc1cecf59.slice/crio-ee65df9a8c3d0007610acee378b398fee47c9218d74cb910edfcf11a7dfa776b WatchSource:0}: Error finding container ee65df9a8c3d0007610acee378b398fee47c9218d74cb910edfcf11a7dfa776b: Status 404 returned error can't find the container with id ee65df9a8c3d0007610acee378b398fee47c9218d74cb910edfcf11a7dfa776b Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.581212 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/549af019-cce7-46bc-a02b-41e1c4e09256-utilities\") pod \"redhat-marketplace-rtc8b\" (UID: \"549af019-cce7-46bc-a02b-41e1c4e09256\") " pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.581416 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqpg9\" (UniqueName: \"kubernetes.io/projected/549af019-cce7-46bc-a02b-41e1c4e09256-kube-api-access-cqpg9\") pod \"redhat-marketplace-rtc8b\" (UID: \"549af019-cce7-46bc-a02b-41e1c4e09256\") " pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.581473 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/549af019-cce7-46bc-a02b-41e1c4e09256-catalog-content\") pod \"redhat-marketplace-rtc8b\" (UID: \"549af019-cce7-46bc-a02b-41e1c4e09256\") " pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.581913 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/549af019-cce7-46bc-a02b-41e1c4e09256-catalog-content\") pod \"redhat-marketplace-rtc8b\" (UID: \"549af019-cce7-46bc-a02b-41e1c4e09256\") " pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.582479 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/549af019-cce7-46bc-a02b-41e1c4e09256-utilities\") pod \"redhat-marketplace-rtc8b\" (UID: \"549af019-cce7-46bc-a02b-41e1c4e09256\") " pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.603098 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqpg9\" (UniqueName: \"kubernetes.io/projected/549af019-cce7-46bc-a02b-41e1c4e09256-kube-api-access-cqpg9\") pod \"redhat-marketplace-rtc8b\" (UID: \"549af019-cce7-46bc-a02b-41e1c4e09256\") " pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.715109 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.946687 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtc8b"] Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.968596 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-26x9k"] Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.971775 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:03:39 crc kubenswrapper[4777]: I1124 17:03:39.975702 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-26x9k"] Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.031486 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.090328 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.095048 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-qz2kt" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.135274 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af02e22-079e-4f81-9f7f-c5b55591387e-catalog-content\") pod \"redhat-operators-26x9k\" (UID: \"3af02e22-079e-4f81-9f7f-c5b55591387e\") " pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.135344 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af02e22-079e-4f81-9f7f-c5b55591387e-utilities\") pod \"redhat-operators-26x9k\" (UID: \"3af02e22-079e-4f81-9f7f-c5b55591387e\") " pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.141931 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjlwm\" (UniqueName: \"kubernetes.io/projected/3af02e22-079e-4f81-9f7f-c5b55591387e-kube-api-access-jjlwm\") pod \"redhat-operators-26x9k\" (UID: \"3af02e22-079e-4f81-9f7f-c5b55591387e\") " pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.157523 4777 patch_prober.go:28] interesting pod/router-default-5444994796-9cm8z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:03:40 crc kubenswrapper[4777]: [-]has-synced failed: reason withheld Nov 24 17:03:40 crc kubenswrapper[4777]: [+]process-running ok Nov 24 17:03:40 crc kubenswrapper[4777]: healthz check failed Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.157584 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9cm8z" podUID="eacac718-a976-4e93-a7af-0414844177fb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.243675 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjlwm\" (UniqueName: \"kubernetes.io/projected/3af02e22-079e-4f81-9f7f-c5b55591387e-kube-api-access-jjlwm\") pod \"redhat-operators-26x9k\" (UID: \"3af02e22-079e-4f81-9f7f-c5b55591387e\") " pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.243741 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af02e22-079e-4f81-9f7f-c5b55591387e-catalog-content\") pod \"redhat-operators-26x9k\" (UID: \"3af02e22-079e-4f81-9f7f-c5b55591387e\") " pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.243771 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af02e22-079e-4f81-9f7f-c5b55591387e-utilities\") pod \"redhat-operators-26x9k\" (UID: \"3af02e22-079e-4f81-9f7f-c5b55591387e\") " pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.244477 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af02e22-079e-4f81-9f7f-c5b55591387e-catalog-content\") pod \"redhat-operators-26x9k\" (UID: \"3af02e22-079e-4f81-9f7f-c5b55591387e\") " pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.244579 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af02e22-079e-4f81-9f7f-c5b55591387e-utilities\") pod \"redhat-operators-26x9k\" (UID: \"3af02e22-079e-4f81-9f7f-c5b55591387e\") " pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.269404 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjlwm\" (UniqueName: \"kubernetes.io/projected/3af02e22-079e-4f81-9f7f-c5b55591387e-kube-api-access-jjlwm\") pod \"redhat-operators-26x9k\" (UID: \"3af02e22-079e-4f81-9f7f-c5b55591387e\") " pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.288571 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.288622 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.307771 4777 patch_prober.go:28] interesting pod/console-f9d7485db-jxbj9 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.307831 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-jxbj9" podUID="10516d2d-106f-4155-ab2c-c0a302ca5cbf" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.368808 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pszfj"] Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.370190 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.380080 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.381906 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pszfj"] Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.485754 4777 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk8pz container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.486136 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vk8pz" podUID="90c90945-24a7-4f17-871c-7358792f0db5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.485764 4777 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk8pz container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.486564 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vk8pz" podUID="90c90945-24a7-4f17-871c-7358792f0db5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.500953 4777 generic.go:334] "Generic (PLEG): container finished" podID="549af019-cce7-46bc-a02b-41e1c4e09256" containerID="bcf14e3d87f861bb6e357707e1c879f17370f22aec6f05f9c9d3c7fd142544b3" exitCode=0 Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.501578 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtc8b" event={"ID":"549af019-cce7-46bc-a02b-41e1c4e09256","Type":"ContainerDied","Data":"bcf14e3d87f861bb6e357707e1c879f17370f22aec6f05f9c9d3c7fd142544b3"} Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.501606 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtc8b" event={"ID":"549af019-cce7-46bc-a02b-41e1c4e09256","Type":"ContainerStarted","Data":"5112763ee0e4be0cbd46dd34a7d02f87b9120baff9c2b453c2140faea4ad2f3a"} Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.505935 4777 generic.go:334] "Generic (PLEG): container finished" podID="e0493374-1b14-4216-bf96-b81a619212e5" containerID="b6a8d9feb9cf6ba3a337820933bb76f15c582214c4d993d88a1caa532fba83f5" exitCode=0 Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.506029 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e0493374-1b14-4216-bf96-b81a619212e5","Type":"ContainerDied","Data":"b6a8d9feb9cf6ba3a337820933bb76f15c582214c4d993d88a1caa532fba83f5"} Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.506046 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e0493374-1b14-4216-bf96-b81a619212e5","Type":"ContainerStarted","Data":"6f0369278739c14cf61eab2f74237a0b7ba3dc68df2276ffe3fe145f1d390fd4"} Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.507525 4777 generic.go:334] "Generic (PLEG): container finished" podID="caacdae7-8c80-4fcb-85de-78dcc1cecf59" containerID="104c59ff8e9c8c45c0d58cc125ef0a8cfbc8f98dffb354be7bde4b7afc82c20c" exitCode=0 Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.508292 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mbg7" event={"ID":"caacdae7-8c80-4fcb-85de-78dcc1cecf59","Type":"ContainerDied","Data":"104c59ff8e9c8c45c0d58cc125ef0a8cfbc8f98dffb354be7bde4b7afc82c20c"} Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.508311 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mbg7" event={"ID":"caacdae7-8c80-4fcb-85de-78dcc1cecf59","Type":"ContainerStarted","Data":"ee65df9a8c3d0007610acee378b398fee47c9218d74cb910edfcf11a7dfa776b"} Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.551515 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-utilities\") pod \"redhat-operators-pszfj\" (UID: \"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e\") " pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.551582 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-catalog-content\") pod \"redhat-operators-pszfj\" (UID: \"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e\") " pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.551648 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6xs7\" (UniqueName: \"kubernetes.io/projected/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-kube-api-access-j6xs7\") pod \"redhat-operators-pszfj\" (UID: \"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e\") " pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.653622 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-utilities\") pod \"redhat-operators-pszfj\" (UID: \"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e\") " pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.653722 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-catalog-content\") pod \"redhat-operators-pszfj\" (UID: \"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e\") " pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.653780 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6xs7\" (UniqueName: \"kubernetes.io/projected/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-kube-api-access-j6xs7\") pod \"redhat-operators-pszfj\" (UID: \"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e\") " pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.655512 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-catalog-content\") pod \"redhat-operators-pszfj\" (UID: \"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e\") " pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.655905 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-utilities\") pod \"redhat-operators-pszfj\" (UID: \"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e\") " pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.675782 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6xs7\" (UniqueName: \"kubernetes.io/projected/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-kube-api-access-j6xs7\") pod \"redhat-operators-pszfj\" (UID: \"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e\") " pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.689892 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.709136 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-26x9k"] Nov 24 17:03:40 crc kubenswrapper[4777]: W1124 17:03:40.734818 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3af02e22_079e_4f81_9f7f_c5b55591387e.slice/crio-7be700959cf6fffed9fe6040d85d381bfdbae7639808b45a40d22209b8152789 WatchSource:0}: Error finding container 7be700959cf6fffed9fe6040d85d381bfdbae7639808b45a40d22209b8152789: Status 404 returned error can't find the container with id 7be700959cf6fffed9fe6040d85d381bfdbae7639808b45a40d22209b8152789 Nov 24 17:03:40 crc kubenswrapper[4777]: I1124 17:03:40.951676 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pszfj"] Nov 24 17:03:40 crc kubenswrapper[4777]: W1124 17:03:40.968331 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a9888ac_755d_4bb9_b5db_07e26a6f1f0e.slice/crio-023a9e8a0aac056e2c519adfab6d6b1442cba0e1912901417e836daf40dbc104 WatchSource:0}: Error finding container 023a9e8a0aac056e2c519adfab6d6b1442cba0e1912901417e836daf40dbc104: Status 404 returned error can't find the container with id 023a9e8a0aac056e2c519adfab6d6b1442cba0e1912901417e836daf40dbc104 Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.131033 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.143063 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.145119 4777 patch_prober.go:28] interesting pod/router-default-5444994796-9cm8z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:03:41 crc kubenswrapper[4777]: [-]has-synced failed: reason withheld Nov 24 17:03:41 crc kubenswrapper[4777]: [+]process-running ok Nov 24 17:03:41 crc kubenswrapper[4777]: healthz check failed Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.145169 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9cm8z" podUID="eacac718-a976-4e93-a7af-0414844177fb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.397750 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.398376 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.403594 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.403880 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.416894 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.545843 4777 generic.go:334] "Generic (PLEG): container finished" podID="3af02e22-079e-4f81-9f7f-c5b55591387e" containerID="5ee96844507a69d62274a029e8bd7bcdb4a4e4a695de59442ef99b7b1dcd9f73" exitCode=0 Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.545913 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-26x9k" event={"ID":"3af02e22-079e-4f81-9f7f-c5b55591387e","Type":"ContainerDied","Data":"5ee96844507a69d62274a029e8bd7bcdb4a4e4a695de59442ef99b7b1dcd9f73"} Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.545939 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-26x9k" event={"ID":"3af02e22-079e-4f81-9f7f-c5b55591387e","Type":"ContainerStarted","Data":"7be700959cf6fffed9fe6040d85d381bfdbae7639808b45a40d22209b8152789"} Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.569763 4777 generic.go:334] "Generic (PLEG): container finished" podID="0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" containerID="7c3619857d5fbabe9964bb5a543c3959ab6d3dc74e19a10c0d8169668abc71e9" exitCode=0 Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.569851 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pszfj" event={"ID":"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e","Type":"ContainerDied","Data":"7c3619857d5fbabe9964bb5a543c3959ab6d3dc74e19a10c0d8169668abc71e9"} Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.569907 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pszfj" event={"ID":"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e","Type":"ContainerStarted","Data":"023a9e8a0aac056e2c519adfab6d6b1442cba0e1912901417e836daf40dbc104"} Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.594293 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/404fb408-cca3-49ee-acb8-69f336a96d0a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"404fb408-cca3-49ee-acb8-69f336a96d0a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.594370 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/404fb408-cca3-49ee-acb8-69f336a96d0a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"404fb408-cca3-49ee-acb8-69f336a96d0a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.695795 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/404fb408-cca3-49ee-acb8-69f336a96d0a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"404fb408-cca3-49ee-acb8-69f336a96d0a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.695962 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/404fb408-cca3-49ee-acb8-69f336a96d0a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"404fb408-cca3-49ee-acb8-69f336a96d0a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.697150 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/404fb408-cca3-49ee-acb8-69f336a96d0a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"404fb408-cca3-49ee-acb8-69f336a96d0a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.716744 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/404fb408-cca3-49ee-acb8-69f336a96d0a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"404fb408-cca3-49ee-acb8-69f336a96d0a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.736754 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:03:41 crc kubenswrapper[4777]: I1124 17:03:41.971708 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:03:42 crc kubenswrapper[4777]: I1124 17:03:42.061172 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 17:03:42 crc kubenswrapper[4777]: I1124 17:03:42.101011 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0493374-1b14-4216-bf96-b81a619212e5-kubelet-dir\") pod \"e0493374-1b14-4216-bf96-b81a619212e5\" (UID: \"e0493374-1b14-4216-bf96-b81a619212e5\") " Nov 24 17:03:42 crc kubenswrapper[4777]: I1124 17:03:42.101111 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e0493374-1b14-4216-bf96-b81a619212e5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e0493374-1b14-4216-bf96-b81a619212e5" (UID: "e0493374-1b14-4216-bf96-b81a619212e5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:03:42 crc kubenswrapper[4777]: I1124 17:03:42.101210 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0493374-1b14-4216-bf96-b81a619212e5-kube-api-access\") pod \"e0493374-1b14-4216-bf96-b81a619212e5\" (UID: \"e0493374-1b14-4216-bf96-b81a619212e5\") " Nov 24 17:03:42 crc kubenswrapper[4777]: I1124 17:03:42.101536 4777 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0493374-1b14-4216-bf96-b81a619212e5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 17:03:42 crc kubenswrapper[4777]: I1124 17:03:42.107190 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0493374-1b14-4216-bf96-b81a619212e5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e0493374-1b14-4216-bf96-b81a619212e5" (UID: "e0493374-1b14-4216-bf96-b81a619212e5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:03:42 crc kubenswrapper[4777]: I1124 17:03:42.147743 4777 patch_prober.go:28] interesting pod/router-default-5444994796-9cm8z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:03:42 crc kubenswrapper[4777]: [-]has-synced failed: reason withheld Nov 24 17:03:42 crc kubenswrapper[4777]: [+]process-running ok Nov 24 17:03:42 crc kubenswrapper[4777]: healthz check failed Nov 24 17:03:42 crc kubenswrapper[4777]: I1124 17:03:42.147815 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9cm8z" podUID="eacac718-a976-4e93-a7af-0414844177fb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:03:42 crc kubenswrapper[4777]: I1124 17:03:42.210615 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0493374-1b14-4216-bf96-b81a619212e5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 17:03:42 crc kubenswrapper[4777]: I1124 17:03:42.588212 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e0493374-1b14-4216-bf96-b81a619212e5","Type":"ContainerDied","Data":"6f0369278739c14cf61eab2f74237a0b7ba3dc68df2276ffe3fe145f1d390fd4"} Nov 24 17:03:42 crc kubenswrapper[4777]: I1124 17:03:42.588252 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f0369278739c14cf61eab2f74237a0b7ba3dc68df2276ffe3fe145f1d390fd4" Nov 24 17:03:42 crc kubenswrapper[4777]: I1124 17:03:42.588220 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 17:03:42 crc kubenswrapper[4777]: I1124 17:03:42.591694 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"404fb408-cca3-49ee-acb8-69f336a96d0a","Type":"ContainerStarted","Data":"7759b40a6a2cb66763c5f1e3ac2da069f9b9c5520b78647aa299dbc69153d122"} Nov 24 17:03:43 crc kubenswrapper[4777]: I1124 17:03:43.145229 4777 patch_prober.go:28] interesting pod/router-default-5444994796-9cm8z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:03:43 crc kubenswrapper[4777]: [-]has-synced failed: reason withheld Nov 24 17:03:43 crc kubenswrapper[4777]: [+]process-running ok Nov 24 17:03:43 crc kubenswrapper[4777]: healthz check failed Nov 24 17:03:43 crc kubenswrapper[4777]: I1124 17:03:43.145284 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9cm8z" podUID="eacac718-a976-4e93-a7af-0414844177fb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:03:43 crc kubenswrapper[4777]: I1124 17:03:43.613033 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"404fb408-cca3-49ee-acb8-69f336a96d0a","Type":"ContainerStarted","Data":"cd9fdceeb4162594f5c0405fd5e844321373edf823db0e053f1910c385b1b20a"} Nov 24 17:03:43 crc kubenswrapper[4777]: I1124 17:03:43.633263 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.633247548 podStartE2EDuration="2.633247548s" podCreationTimestamp="2025-11-24 17:03:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:03:43.630957214 +0000 UTC m=+181.790152263" watchObservedRunningTime="2025-11-24 17:03:43.633247548 +0000 UTC m=+181.792442587" Nov 24 17:03:44 crc kubenswrapper[4777]: I1124 17:03:44.150854 4777 patch_prober.go:28] interesting pod/router-default-5444994796-9cm8z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:03:44 crc kubenswrapper[4777]: [-]has-synced failed: reason withheld Nov 24 17:03:44 crc kubenswrapper[4777]: [+]process-running ok Nov 24 17:03:44 crc kubenswrapper[4777]: healthz check failed Nov 24 17:03:44 crc kubenswrapper[4777]: I1124 17:03:44.150900 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9cm8z" podUID="eacac718-a976-4e93-a7af-0414844177fb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:03:44 crc kubenswrapper[4777]: I1124 17:03:44.269916 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:03:44 crc kubenswrapper[4777]: I1124 17:03:44.269995 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:03:44 crc kubenswrapper[4777]: I1124 17:03:44.631474 4777 generic.go:334] "Generic (PLEG): container finished" podID="404fb408-cca3-49ee-acb8-69f336a96d0a" containerID="cd9fdceeb4162594f5c0405fd5e844321373edf823db0e053f1910c385b1b20a" exitCode=0 Nov 24 17:03:44 crc kubenswrapper[4777]: I1124 17:03:44.631733 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"404fb408-cca3-49ee-acb8-69f336a96d0a","Type":"ContainerDied","Data":"cd9fdceeb4162594f5c0405fd5e844321373edf823db0e053f1910c385b1b20a"} Nov 24 17:03:45 crc kubenswrapper[4777]: I1124 17:03:45.148484 4777 patch_prober.go:28] interesting pod/router-default-5444994796-9cm8z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:03:45 crc kubenswrapper[4777]: [-]has-synced failed: reason withheld Nov 24 17:03:45 crc kubenswrapper[4777]: [+]process-running ok Nov 24 17:03:45 crc kubenswrapper[4777]: healthz check failed Nov 24 17:03:45 crc kubenswrapper[4777]: I1124 17:03:45.148535 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9cm8z" podUID="eacac718-a976-4e93-a7af-0414844177fb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:03:45 crc kubenswrapper[4777]: I1124 17:03:45.861471 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-wbx6n" Nov 24 17:03:46 crc kubenswrapper[4777]: I1124 17:03:46.144958 4777 patch_prober.go:28] interesting pod/router-default-5444994796-9cm8z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 17:03:46 crc kubenswrapper[4777]: [-]has-synced failed: reason withheld Nov 24 17:03:46 crc kubenswrapper[4777]: [+]process-running ok Nov 24 17:03:46 crc kubenswrapper[4777]: healthz check failed Nov 24 17:03:46 crc kubenswrapper[4777]: I1124 17:03:46.145248 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9cm8z" podUID="eacac718-a976-4e93-a7af-0414844177fb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:03:47 crc kubenswrapper[4777]: I1124 17:03:47.145476 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:47 crc kubenswrapper[4777]: I1124 17:03:47.149334 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-9cm8z" Nov 24 17:03:50 crc kubenswrapper[4777]: I1124 17:03:50.300946 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:50 crc kubenswrapper[4777]: I1124 17:03:50.305823 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:03:50 crc kubenswrapper[4777]: I1124 17:03:50.484276 4777 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk8pz container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Nov 24 17:03:50 crc kubenswrapper[4777]: I1124 17:03:50.484358 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vk8pz" podUID="90c90945-24a7-4f17-871c-7358792f0db5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Nov 24 17:03:50 crc kubenswrapper[4777]: I1124 17:03:50.484524 4777 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk8pz container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Nov 24 17:03:50 crc kubenswrapper[4777]: I1124 17:03:50.484586 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vk8pz" podUID="90c90945-24a7-4f17-871c-7358792f0db5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Nov 24 17:03:50 crc kubenswrapper[4777]: I1124 17:03:50.747661 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:03:50 crc kubenswrapper[4777]: I1124 17:03:50.872463 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/404fb408-cca3-49ee-acb8-69f336a96d0a-kubelet-dir\") pod \"404fb408-cca3-49ee-acb8-69f336a96d0a\" (UID: \"404fb408-cca3-49ee-acb8-69f336a96d0a\") " Nov 24 17:03:50 crc kubenswrapper[4777]: I1124 17:03:50.872619 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/404fb408-cca3-49ee-acb8-69f336a96d0a-kube-api-access\") pod \"404fb408-cca3-49ee-acb8-69f336a96d0a\" (UID: \"404fb408-cca3-49ee-acb8-69f336a96d0a\") " Nov 24 17:03:50 crc kubenswrapper[4777]: I1124 17:03:50.872651 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/404fb408-cca3-49ee-acb8-69f336a96d0a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "404fb408-cca3-49ee-acb8-69f336a96d0a" (UID: "404fb408-cca3-49ee-acb8-69f336a96d0a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:03:50 crc kubenswrapper[4777]: I1124 17:03:50.873607 4777 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/404fb408-cca3-49ee-acb8-69f336a96d0a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 17:03:50 crc kubenswrapper[4777]: I1124 17:03:50.880182 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/404fb408-cca3-49ee-acb8-69f336a96d0a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "404fb408-cca3-49ee-acb8-69f336a96d0a" (UID: "404fb408-cca3-49ee-acb8-69f336a96d0a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:03:50 crc kubenswrapper[4777]: I1124 17:03:50.974802 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/404fb408-cca3-49ee-acb8-69f336a96d0a-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 17:03:51 crc kubenswrapper[4777]: I1124 17:03:51.678331 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"404fb408-cca3-49ee-acb8-69f336a96d0a","Type":"ContainerDied","Data":"7759b40a6a2cb66763c5f1e3ac2da069f9b9c5520b78647aa299dbc69153d122"} Nov 24 17:03:51 crc kubenswrapper[4777]: I1124 17:03:51.678817 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7759b40a6a2cb66763c5f1e3ac2da069f9b9c5520b78647aa299dbc69153d122" Nov 24 17:03:51 crc kubenswrapper[4777]: I1124 17:03:51.678505 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 17:03:58 crc kubenswrapper[4777]: I1124 17:03:58.041882 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:04:00 crc kubenswrapper[4777]: I1124 17:04:00.490104 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-vk8pz" Nov 24 17:04:07 crc kubenswrapper[4777]: E1124 17:04:07.902386 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 24 17:04:07 crc kubenswrapper[4777]: E1124 17:04:07.903040 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b6qnd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-tr7gm_openshift-marketplace(6203d413-faf0-4476-9a91-345ba4f1772f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 17:04:07 crc kubenswrapper[4777]: E1124 17:04:07.904308 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-tr7gm" podUID="6203d413-faf0-4476-9a91-345ba4f1772f" Nov 24 17:04:09 crc kubenswrapper[4777]: E1124 17:04:09.639888 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 24 17:04:09 crc kubenswrapper[4777]: E1124 17:04:09.640464 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gg67t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-hqk2w_openshift-marketplace(48ca8fd0-3e8b-4df1-84c6-01104522f616): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 17:04:09 crc kubenswrapper[4777]: E1124 17:04:09.641748 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-hqk2w" podUID="48ca8fd0-3e8b-4df1-84c6-01104522f616" Nov 24 17:04:09 crc kubenswrapper[4777]: E1124 17:04:09.652436 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 24 17:04:09 crc kubenswrapper[4777]: E1124 17:04:09.652673 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cqpg9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-rtc8b_openshift-marketplace(549af019-cce7-46bc-a02b-41e1c4e09256): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 17:04:09 crc kubenswrapper[4777]: E1124 17:04:09.653958 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-rtc8b" podUID="549af019-cce7-46bc-a02b-41e1c4e09256" Nov 24 17:04:11 crc kubenswrapper[4777]: I1124 17:04:11.279057 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cqdzb" Nov 24 17:04:12 crc kubenswrapper[4777]: E1124 17:04:12.750585 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-hqk2w" podUID="48ca8fd0-3e8b-4df1-84c6-01104522f616" Nov 24 17:04:12 crc kubenswrapper[4777]: E1124 17:04:12.750749 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-rtc8b" podUID="549af019-cce7-46bc-a02b-41e1c4e09256" Nov 24 17:04:12 crc kubenswrapper[4777]: E1124 17:04:12.750798 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tr7gm" podUID="6203d413-faf0-4476-9a91-345ba4f1772f" Nov 24 17:04:12 crc kubenswrapper[4777]: E1124 17:04:12.855031 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 24 17:04:12 crc kubenswrapper[4777]: E1124 17:04:12.855149 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j6xs7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-pszfj_openshift-marketplace(0a9888ac-755d-4bb9-b5db-07e26a6f1f0e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 17:04:12 crc kubenswrapper[4777]: E1124 17:04:12.856495 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-pszfj" podUID="0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" Nov 24 17:04:13 crc kubenswrapper[4777]: I1124 17:04:13.813944 4777 generic.go:334] "Generic (PLEG): container finished" podID="caacdae7-8c80-4fcb-85de-78dcc1cecf59" containerID="45f4b0a7abe4dc38fd70eea11f2e3b2240cc21d3987197911ec0eaac43c49574" exitCode=0 Nov 24 17:04:13 crc kubenswrapper[4777]: I1124 17:04:13.814157 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mbg7" event={"ID":"caacdae7-8c80-4fcb-85de-78dcc1cecf59","Type":"ContainerDied","Data":"45f4b0a7abe4dc38fd70eea11f2e3b2240cc21d3987197911ec0eaac43c49574"} Nov 24 17:04:13 crc kubenswrapper[4777]: I1124 17:04:13.818666 4777 generic.go:334] "Generic (PLEG): container finished" podID="3af02e22-079e-4f81-9f7f-c5b55591387e" containerID="f16a0f8f2ee0d857a2639480c01badf387ba3ad1d6521fd9db5d38a82d0e3606" exitCode=0 Nov 24 17:04:13 crc kubenswrapper[4777]: I1124 17:04:13.818779 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-26x9k" event={"ID":"3af02e22-079e-4f81-9f7f-c5b55591387e","Type":"ContainerDied","Data":"f16a0f8f2ee0d857a2639480c01badf387ba3ad1d6521fd9db5d38a82d0e3606"} Nov 24 17:04:13 crc kubenswrapper[4777]: I1124 17:04:13.827676 4777 generic.go:334] "Generic (PLEG): container finished" podID="eebaa807-fbd5-40d7-ac22-02523cbc4fc2" containerID="ad95be6a85802aa2a8b4a2936e69b5d69a99f1a9695940581e9961de8fe30d6b" exitCode=0 Nov 24 17:04:13 crc kubenswrapper[4777]: I1124 17:04:13.827760 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmfnz" event={"ID":"eebaa807-fbd5-40d7-ac22-02523cbc4fc2","Type":"ContainerDied","Data":"ad95be6a85802aa2a8b4a2936e69b5d69a99f1a9695940581e9961de8fe30d6b"} Nov 24 17:04:13 crc kubenswrapper[4777]: I1124 17:04:13.836563 4777 generic.go:334] "Generic (PLEG): container finished" podID="b93bcc12-1487-4c0f-9da2-fab9c61df5c9" containerID="77413ac0483f9186775c6ac946883282de8436766b13e3fe00a3906d741ddd07" exitCode=0 Nov 24 17:04:13 crc kubenswrapper[4777]: I1124 17:04:13.838624 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-shz64" event={"ID":"b93bcc12-1487-4c0f-9da2-fab9c61df5c9","Type":"ContainerDied","Data":"77413ac0483f9186775c6ac946883282de8436766b13e3fe00a3906d741ddd07"} Nov 24 17:04:13 crc kubenswrapper[4777]: E1124 17:04:13.838811 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-pszfj" podUID="0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" Nov 24 17:04:14 crc kubenswrapper[4777]: I1124 17:04:14.270209 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:04:14 crc kubenswrapper[4777]: I1124 17:04:14.270729 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:04:14 crc kubenswrapper[4777]: I1124 17:04:14.270802 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:04:14 crc kubenswrapper[4777]: I1124 17:04:14.271714 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:04:14 crc kubenswrapper[4777]: I1124 17:04:14.271884 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3" gracePeriod=600 Nov 24 17:04:14 crc kubenswrapper[4777]: I1124 17:04:14.861855 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-shz64" event={"ID":"b93bcc12-1487-4c0f-9da2-fab9c61df5c9","Type":"ContainerStarted","Data":"bbd16d4ddad153f2a9d67eb28915908ece9b183567ba4018777fed0cacd79014"} Nov 24 17:04:14 crc kubenswrapper[4777]: I1124 17:04:14.870136 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3" exitCode=0 Nov 24 17:04:14 crc kubenswrapper[4777]: I1124 17:04:14.870201 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3"} Nov 24 17:04:14 crc kubenswrapper[4777]: I1124 17:04:14.870228 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"e035704384e89657cd1567ffda73b162c76374285cb18e948ce33a5511ac19b6"} Nov 24 17:04:14 crc kubenswrapper[4777]: I1124 17:04:14.873727 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mbg7" event={"ID":"caacdae7-8c80-4fcb-85de-78dcc1cecf59","Type":"ContainerStarted","Data":"9e115b1137cd5fa9f36d9d39f4ac62d1cf331b53a68ed9932ac520442b78f452"} Nov 24 17:04:14 crc kubenswrapper[4777]: I1124 17:04:14.883101 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-26x9k" event={"ID":"3af02e22-079e-4f81-9f7f-c5b55591387e","Type":"ContainerStarted","Data":"a8a8ef411766bb3846abc92b8692f7065a3f509759ebdaf136f9fbac914a9d1d"} Nov 24 17:04:14 crc kubenswrapper[4777]: I1124 17:04:14.883240 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-shz64" podStartSLOduration=2.991689267 podStartE2EDuration="38.883222401s" podCreationTimestamp="2025-11-24 17:03:36 +0000 UTC" firstStartedPulling="2025-11-24 17:03:38.444166253 +0000 UTC m=+176.603361302" lastFinishedPulling="2025-11-24 17:04:14.335699377 +0000 UTC m=+212.494894436" observedRunningTime="2025-11-24 17:04:14.879138371 +0000 UTC m=+213.038333460" watchObservedRunningTime="2025-11-24 17:04:14.883222401 +0000 UTC m=+213.042417450" Nov 24 17:04:14 crc kubenswrapper[4777]: I1124 17:04:14.897687 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmfnz" event={"ID":"eebaa807-fbd5-40d7-ac22-02523cbc4fc2","Type":"ContainerStarted","Data":"f200549fe9ed4eb4e5561a3b568ff56149a7f5f07f27ef9f5aef5dcb609e43e9"} Nov 24 17:04:14 crc kubenswrapper[4777]: I1124 17:04:14.923762 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9mbg7" podStartSLOduration=3.177604608 podStartE2EDuration="36.923740667s" podCreationTimestamp="2025-11-24 17:03:38 +0000 UTC" firstStartedPulling="2025-11-24 17:03:40.509229346 +0000 UTC m=+178.668424395" lastFinishedPulling="2025-11-24 17:04:14.255365395 +0000 UTC m=+212.414560454" observedRunningTime="2025-11-24 17:04:14.919622575 +0000 UTC m=+213.078817644" watchObservedRunningTime="2025-11-24 17:04:14.923740667 +0000 UTC m=+213.082935726" Nov 24 17:04:14 crc kubenswrapper[4777]: I1124 17:04:14.945680 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wmfnz" podStartSLOduration=2.013563477 podStartE2EDuration="37.945661159s" podCreationTimestamp="2025-11-24 17:03:37 +0000 UTC" firstStartedPulling="2025-11-24 17:03:38.432580739 +0000 UTC m=+176.591775788" lastFinishedPulling="2025-11-24 17:04:14.364678411 +0000 UTC m=+212.523873470" observedRunningTime="2025-11-24 17:04:14.943959483 +0000 UTC m=+213.103154522" watchObservedRunningTime="2025-11-24 17:04:14.945661159 +0000 UTC m=+213.104856208" Nov 24 17:04:17 crc kubenswrapper[4777]: I1124 17:04:17.135737 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:04:17 crc kubenswrapper[4777]: I1124 17:04:17.137083 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:04:17 crc kubenswrapper[4777]: I1124 17:04:17.318516 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:04:17 crc kubenswrapper[4777]: I1124 17:04:17.346025 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-26x9k" podStartSLOduration=5.656958368 podStartE2EDuration="38.346009658s" podCreationTimestamp="2025-11-24 17:03:39 +0000 UTC" firstStartedPulling="2025-11-24 17:03:41.558157865 +0000 UTC m=+179.717352914" lastFinishedPulling="2025-11-24 17:04:14.247209145 +0000 UTC m=+212.406404204" observedRunningTime="2025-11-24 17:04:14.960043138 +0000 UTC m=+213.119238187" watchObservedRunningTime="2025-11-24 17:04:17.346009658 +0000 UTC m=+215.505204707" Nov 24 17:04:17 crc kubenswrapper[4777]: I1124 17:04:17.524505 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:04:17 crc kubenswrapper[4777]: I1124 17:04:17.525694 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:04:17 crc kubenswrapper[4777]: I1124 17:04:17.572852 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:04:19 crc kubenswrapper[4777]: I1124 17:04:19.315693 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:04:19 crc kubenswrapper[4777]: I1124 17:04:19.316066 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:04:19 crc kubenswrapper[4777]: I1124 17:04:19.325278 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7hwq6"] Nov 24 17:04:19 crc kubenswrapper[4777]: I1124 17:04:19.383420 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:04:20 crc kubenswrapper[4777]: I1124 17:04:20.286556 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:04:20 crc kubenswrapper[4777]: I1124 17:04:20.381329 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:04:20 crc kubenswrapper[4777]: I1124 17:04:20.381392 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:04:20 crc kubenswrapper[4777]: I1124 17:04:20.426871 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:04:21 crc kubenswrapper[4777]: I1124 17:04:21.278151 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:04:27 crc kubenswrapper[4777]: I1124 17:04:27.174442 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:04:27 crc kubenswrapper[4777]: I1124 17:04:27.567825 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:04:28 crc kubenswrapper[4777]: I1124 17:04:28.286753 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr7gm" event={"ID":"6203d413-faf0-4476-9a91-345ba4f1772f","Type":"ContainerStarted","Data":"322b76ed6de4ddbafd27e4249776fe3ab148c238d62681752038a73991ff9f9c"} Nov 24 17:04:28 crc kubenswrapper[4777]: I1124 17:04:28.289013 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pszfj" event={"ID":"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e","Type":"ContainerStarted","Data":"2f726df1ec7b12fd184201d6662815d8deeea50748906599a1f5f7c9962bd77d"} Nov 24 17:04:28 crc kubenswrapper[4777]: I1124 17:04:28.291986 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqk2w" event={"ID":"48ca8fd0-3e8b-4df1-84c6-01104522f616","Type":"ContainerStarted","Data":"e48817a026f61baf17c645029ceb707e359ad6784e666a118679960bdc6eccaf"} Nov 24 17:04:29 crc kubenswrapper[4777]: I1124 17:04:29.302393 4777 generic.go:334] "Generic (PLEG): container finished" podID="0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" containerID="2f726df1ec7b12fd184201d6662815d8deeea50748906599a1f5f7c9962bd77d" exitCode=0 Nov 24 17:04:29 crc kubenswrapper[4777]: I1124 17:04:29.302464 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pszfj" event={"ID":"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e","Type":"ContainerDied","Data":"2f726df1ec7b12fd184201d6662815d8deeea50748906599a1f5f7c9962bd77d"} Nov 24 17:04:29 crc kubenswrapper[4777]: I1124 17:04:29.305257 4777 generic.go:334] "Generic (PLEG): container finished" podID="48ca8fd0-3e8b-4df1-84c6-01104522f616" containerID="e48817a026f61baf17c645029ceb707e359ad6784e666a118679960bdc6eccaf" exitCode=0 Nov 24 17:04:29 crc kubenswrapper[4777]: I1124 17:04:29.305320 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqk2w" event={"ID":"48ca8fd0-3e8b-4df1-84c6-01104522f616","Type":"ContainerDied","Data":"e48817a026f61baf17c645029ceb707e359ad6784e666a118679960bdc6eccaf"} Nov 24 17:04:29 crc kubenswrapper[4777]: I1124 17:04:29.308589 4777 generic.go:334] "Generic (PLEG): container finished" podID="6203d413-faf0-4476-9a91-345ba4f1772f" containerID="322b76ed6de4ddbafd27e4249776fe3ab148c238d62681752038a73991ff9f9c" exitCode=0 Nov 24 17:04:29 crc kubenswrapper[4777]: I1124 17:04:29.308665 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr7gm" event={"ID":"6203d413-faf0-4476-9a91-345ba4f1772f","Type":"ContainerDied","Data":"322b76ed6de4ddbafd27e4249776fe3ab148c238d62681752038a73991ff9f9c"} Nov 24 17:04:29 crc kubenswrapper[4777]: I1124 17:04:29.311775 4777 generic.go:334] "Generic (PLEG): container finished" podID="549af019-cce7-46bc-a02b-41e1c4e09256" containerID="e6c96623697ef8154219db817e1c8f32fedf0f22a2c9ba97d276601d4b43bfcb" exitCode=0 Nov 24 17:04:29 crc kubenswrapper[4777]: I1124 17:04:29.311820 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtc8b" event={"ID":"549af019-cce7-46bc-a02b-41e1c4e09256","Type":"ContainerDied","Data":"e6c96623697ef8154219db817e1c8f32fedf0f22a2c9ba97d276601d4b43bfcb"} Nov 24 17:04:30 crc kubenswrapper[4777]: I1124 17:04:30.319906 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtc8b" event={"ID":"549af019-cce7-46bc-a02b-41e1c4e09256","Type":"ContainerStarted","Data":"080938aaad95886e016a76524d3b1870ad7e822438b378443454a501f32d80f5"} Nov 24 17:04:30 crc kubenswrapper[4777]: I1124 17:04:30.322995 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pszfj" event={"ID":"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e","Type":"ContainerStarted","Data":"0e5419d3f1d8fea39c8aba18c33a9f73e58130a7e761c0a2e348a322013f3b62"} Nov 24 17:04:30 crc kubenswrapper[4777]: I1124 17:04:30.325354 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqk2w" event={"ID":"48ca8fd0-3e8b-4df1-84c6-01104522f616","Type":"ContainerStarted","Data":"30ce0fb41d57b8d9015f46071889f7699381f8a114a7a1f73cea8d12533132ab"} Nov 24 17:04:30 crc kubenswrapper[4777]: I1124 17:04:30.328712 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr7gm" event={"ID":"6203d413-faf0-4476-9a91-345ba4f1772f","Type":"ContainerStarted","Data":"ecc95bdd713cd10e0c52925e7fedd4aec58c072eb482b738a1e2c5c81300be65"} Nov 24 17:04:30 crc kubenswrapper[4777]: I1124 17:04:30.347913 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rtc8b" podStartSLOduration=2.134536697 podStartE2EDuration="51.347891138s" podCreationTimestamp="2025-11-24 17:03:39 +0000 UTC" firstStartedPulling="2025-11-24 17:03:40.502541989 +0000 UTC m=+178.661737038" lastFinishedPulling="2025-11-24 17:04:29.71589643 +0000 UTC m=+227.875091479" observedRunningTime="2025-11-24 17:04:30.346589202 +0000 UTC m=+228.505784271" watchObservedRunningTime="2025-11-24 17:04:30.347891138 +0000 UTC m=+228.507086197" Nov 24 17:04:30 crc kubenswrapper[4777]: I1124 17:04:30.369461 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tr7gm" podStartSLOduration=2.990108828 podStartE2EDuration="54.36944536s" podCreationTimestamp="2025-11-24 17:03:36 +0000 UTC" firstStartedPulling="2025-11-24 17:03:38.438489184 +0000 UTC m=+176.597684233" lastFinishedPulling="2025-11-24 17:04:29.817825716 +0000 UTC m=+227.977020765" observedRunningTime="2025-11-24 17:04:30.366008357 +0000 UTC m=+228.525203406" watchObservedRunningTime="2025-11-24 17:04:30.36944536 +0000 UTC m=+228.528640409" Nov 24 17:04:30 crc kubenswrapper[4777]: I1124 17:04:30.387511 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hqk2w" podStartSLOduration=3.123856497 podStartE2EDuration="53.387497108s" podCreationTimestamp="2025-11-24 17:03:37 +0000 UTC" firstStartedPulling="2025-11-24 17:03:39.48263901 +0000 UTC m=+177.641834059" lastFinishedPulling="2025-11-24 17:04:29.746279611 +0000 UTC m=+227.905474670" observedRunningTime="2025-11-24 17:04:30.38717043 +0000 UTC m=+228.546365479" watchObservedRunningTime="2025-11-24 17:04:30.387497108 +0000 UTC m=+228.546692147" Nov 24 17:04:30 crc kubenswrapper[4777]: I1124 17:04:30.408895 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pszfj" podStartSLOduration=2.1734692349999998 podStartE2EDuration="50.408881097s" podCreationTimestamp="2025-11-24 17:03:40 +0000 UTC" firstStartedPulling="2025-11-24 17:03:41.572993679 +0000 UTC m=+179.732188728" lastFinishedPulling="2025-11-24 17:04:29.808405551 +0000 UTC m=+227.967600590" observedRunningTime="2025-11-24 17:04:30.40307913 +0000 UTC m=+228.562274179" watchObservedRunningTime="2025-11-24 17:04:30.408881097 +0000 UTC m=+228.568076146" Nov 24 17:04:30 crc kubenswrapper[4777]: I1124 17:04:30.691046 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:04:30 crc kubenswrapper[4777]: I1124 17:04:30.691099 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:04:31 crc kubenswrapper[4777]: I1124 17:04:31.583904 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wmfnz"] Nov 24 17:04:31 crc kubenswrapper[4777]: I1124 17:04:31.584164 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wmfnz" podUID="eebaa807-fbd5-40d7-ac22-02523cbc4fc2" containerName="registry-server" containerID="cri-o://f200549fe9ed4eb4e5561a3b568ff56149a7f5f07f27ef9f5aef5dcb609e43e9" gracePeriod=2 Nov 24 17:04:31 crc kubenswrapper[4777]: I1124 17:04:31.735496 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pszfj" podUID="0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" containerName="registry-server" probeResult="failure" output=< Nov 24 17:04:31 crc kubenswrapper[4777]: timeout: failed to connect service ":50051" within 1s Nov 24 17:04:31 crc kubenswrapper[4777]: > Nov 24 17:04:31 crc kubenswrapper[4777]: I1124 17:04:31.935370 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.115077 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpdcz\" (UniqueName: \"kubernetes.io/projected/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-kube-api-access-fpdcz\") pod \"eebaa807-fbd5-40d7-ac22-02523cbc4fc2\" (UID: \"eebaa807-fbd5-40d7-ac22-02523cbc4fc2\") " Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.115162 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-catalog-content\") pod \"eebaa807-fbd5-40d7-ac22-02523cbc4fc2\" (UID: \"eebaa807-fbd5-40d7-ac22-02523cbc4fc2\") " Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.115246 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-utilities\") pod \"eebaa807-fbd5-40d7-ac22-02523cbc4fc2\" (UID: \"eebaa807-fbd5-40d7-ac22-02523cbc4fc2\") " Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.118147 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-utilities" (OuterVolumeSpecName: "utilities") pod "eebaa807-fbd5-40d7-ac22-02523cbc4fc2" (UID: "eebaa807-fbd5-40d7-ac22-02523cbc4fc2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.122084 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-kube-api-access-fpdcz" (OuterVolumeSpecName: "kube-api-access-fpdcz") pod "eebaa807-fbd5-40d7-ac22-02523cbc4fc2" (UID: "eebaa807-fbd5-40d7-ac22-02523cbc4fc2"). InnerVolumeSpecName "kube-api-access-fpdcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.165674 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eebaa807-fbd5-40d7-ac22-02523cbc4fc2" (UID: "eebaa807-fbd5-40d7-ac22-02523cbc4fc2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.216822 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpdcz\" (UniqueName: \"kubernetes.io/projected/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-kube-api-access-fpdcz\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.216853 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.216866 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eebaa807-fbd5-40d7-ac22-02523cbc4fc2-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.348899 4777 generic.go:334] "Generic (PLEG): container finished" podID="eebaa807-fbd5-40d7-ac22-02523cbc4fc2" containerID="f200549fe9ed4eb4e5561a3b568ff56149a7f5f07f27ef9f5aef5dcb609e43e9" exitCode=0 Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.349037 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmfnz" event={"ID":"eebaa807-fbd5-40d7-ac22-02523cbc4fc2","Type":"ContainerDied","Data":"f200549fe9ed4eb4e5561a3b568ff56149a7f5f07f27ef9f5aef5dcb609e43e9"} Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.349086 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wmfnz" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.349117 4777 scope.go:117] "RemoveContainer" containerID="f200549fe9ed4eb4e5561a3b568ff56149a7f5f07f27ef9f5aef5dcb609e43e9" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.349094 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmfnz" event={"ID":"eebaa807-fbd5-40d7-ac22-02523cbc4fc2","Type":"ContainerDied","Data":"ae4299f1cc5dbfc5f9a6757920eb49cf791c404b6ba976e8fa8507cb12492137"} Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.371679 4777 scope.go:117] "RemoveContainer" containerID="ad95be6a85802aa2a8b4a2936e69b5d69a99f1a9695940581e9961de8fe30d6b" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.393108 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wmfnz"] Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.394835 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wmfnz"] Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.395821 4777 scope.go:117] "RemoveContainer" containerID="f1698cab81e104818a14e4285bcfa695c045c6f501c3461def680fe9a145c478" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.424951 4777 scope.go:117] "RemoveContainer" containerID="f200549fe9ed4eb4e5561a3b568ff56149a7f5f07f27ef9f5aef5dcb609e43e9" Nov 24 17:04:32 crc kubenswrapper[4777]: E1124 17:04:32.425513 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f200549fe9ed4eb4e5561a3b568ff56149a7f5f07f27ef9f5aef5dcb609e43e9\": container with ID starting with f200549fe9ed4eb4e5561a3b568ff56149a7f5f07f27ef9f5aef5dcb609e43e9 not found: ID does not exist" containerID="f200549fe9ed4eb4e5561a3b568ff56149a7f5f07f27ef9f5aef5dcb609e43e9" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.425556 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f200549fe9ed4eb4e5561a3b568ff56149a7f5f07f27ef9f5aef5dcb609e43e9"} err="failed to get container status \"f200549fe9ed4eb4e5561a3b568ff56149a7f5f07f27ef9f5aef5dcb609e43e9\": rpc error: code = NotFound desc = could not find container \"f200549fe9ed4eb4e5561a3b568ff56149a7f5f07f27ef9f5aef5dcb609e43e9\": container with ID starting with f200549fe9ed4eb4e5561a3b568ff56149a7f5f07f27ef9f5aef5dcb609e43e9 not found: ID does not exist" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.425585 4777 scope.go:117] "RemoveContainer" containerID="ad95be6a85802aa2a8b4a2936e69b5d69a99f1a9695940581e9961de8fe30d6b" Nov 24 17:04:32 crc kubenswrapper[4777]: E1124 17:04:32.426081 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad95be6a85802aa2a8b4a2936e69b5d69a99f1a9695940581e9961de8fe30d6b\": container with ID starting with ad95be6a85802aa2a8b4a2936e69b5d69a99f1a9695940581e9961de8fe30d6b not found: ID does not exist" containerID="ad95be6a85802aa2a8b4a2936e69b5d69a99f1a9695940581e9961de8fe30d6b" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.426113 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad95be6a85802aa2a8b4a2936e69b5d69a99f1a9695940581e9961de8fe30d6b"} err="failed to get container status \"ad95be6a85802aa2a8b4a2936e69b5d69a99f1a9695940581e9961de8fe30d6b\": rpc error: code = NotFound desc = could not find container \"ad95be6a85802aa2a8b4a2936e69b5d69a99f1a9695940581e9961de8fe30d6b\": container with ID starting with ad95be6a85802aa2a8b4a2936e69b5d69a99f1a9695940581e9961de8fe30d6b not found: ID does not exist" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.426135 4777 scope.go:117] "RemoveContainer" containerID="f1698cab81e104818a14e4285bcfa695c045c6f501c3461def680fe9a145c478" Nov 24 17:04:32 crc kubenswrapper[4777]: E1124 17:04:32.426636 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1698cab81e104818a14e4285bcfa695c045c6f501c3461def680fe9a145c478\": container with ID starting with f1698cab81e104818a14e4285bcfa695c045c6f501c3461def680fe9a145c478 not found: ID does not exist" containerID="f1698cab81e104818a14e4285bcfa695c045c6f501c3461def680fe9a145c478" Nov 24 17:04:32 crc kubenswrapper[4777]: I1124 17:04:32.426697 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1698cab81e104818a14e4285bcfa695c045c6f501c3461def680fe9a145c478"} err="failed to get container status \"f1698cab81e104818a14e4285bcfa695c045c6f501c3461def680fe9a145c478\": rpc error: code = NotFound desc = could not find container \"f1698cab81e104818a14e4285bcfa695c045c6f501c3461def680fe9a145c478\": container with ID starting with f1698cab81e104818a14e4285bcfa695c045c6f501c3461def680fe9a145c478 not found: ID does not exist" Nov 24 17:04:33 crc kubenswrapper[4777]: I1124 17:04:33.255609 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eebaa807-fbd5-40d7-ac22-02523cbc4fc2" path="/var/lib/kubelet/pods/eebaa807-fbd5-40d7-ac22-02523cbc4fc2/volumes" Nov 24 17:04:37 crc kubenswrapper[4777]: I1124 17:04:37.303438 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:04:37 crc kubenswrapper[4777]: I1124 17:04:37.303942 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:04:37 crc kubenswrapper[4777]: I1124 17:04:37.385788 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:04:37 crc kubenswrapper[4777]: I1124 17:04:37.455428 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:04:37 crc kubenswrapper[4777]: I1124 17:04:37.731322 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:04:37 crc kubenswrapper[4777]: I1124 17:04:37.731373 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:04:37 crc kubenswrapper[4777]: I1124 17:04:37.788885 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:04:38 crc kubenswrapper[4777]: I1124 17:04:38.456680 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:04:39 crc kubenswrapper[4777]: I1124 17:04:39.410225 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hqk2w"] Nov 24 17:04:39 crc kubenswrapper[4777]: I1124 17:04:39.715732 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:04:39 crc kubenswrapper[4777]: I1124 17:04:39.715857 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:04:39 crc kubenswrapper[4777]: I1124 17:04:39.784726 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:04:40 crc kubenswrapper[4777]: I1124 17:04:40.408448 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hqk2w" podUID="48ca8fd0-3e8b-4df1-84c6-01104522f616" containerName="registry-server" containerID="cri-o://30ce0fb41d57b8d9015f46071889f7699381f8a114a7a1f73cea8d12533132ab" gracePeriod=2 Nov 24 17:04:40 crc kubenswrapper[4777]: I1124 17:04:40.485273 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:04:40 crc kubenswrapper[4777]: I1124 17:04:40.759141 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:04:40 crc kubenswrapper[4777]: I1124 17:04:40.832381 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.205638 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.355246 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48ca8fd0-3e8b-4df1-84c6-01104522f616-catalog-content\") pod \"48ca8fd0-3e8b-4df1-84c6-01104522f616\" (UID: \"48ca8fd0-3e8b-4df1-84c6-01104522f616\") " Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.355303 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48ca8fd0-3e8b-4df1-84c6-01104522f616-utilities\") pod \"48ca8fd0-3e8b-4df1-84c6-01104522f616\" (UID: \"48ca8fd0-3e8b-4df1-84c6-01104522f616\") " Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.355350 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gg67t\" (UniqueName: \"kubernetes.io/projected/48ca8fd0-3e8b-4df1-84c6-01104522f616-kube-api-access-gg67t\") pod \"48ca8fd0-3e8b-4df1-84c6-01104522f616\" (UID: \"48ca8fd0-3e8b-4df1-84c6-01104522f616\") " Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.357380 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48ca8fd0-3e8b-4df1-84c6-01104522f616-utilities" (OuterVolumeSpecName: "utilities") pod "48ca8fd0-3e8b-4df1-84c6-01104522f616" (UID: "48ca8fd0-3e8b-4df1-84c6-01104522f616"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.364143 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48ca8fd0-3e8b-4df1-84c6-01104522f616-kube-api-access-gg67t" (OuterVolumeSpecName: "kube-api-access-gg67t") pod "48ca8fd0-3e8b-4df1-84c6-01104522f616" (UID: "48ca8fd0-3e8b-4df1-84c6-01104522f616"). InnerVolumeSpecName "kube-api-access-gg67t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.422731 4777 generic.go:334] "Generic (PLEG): container finished" podID="48ca8fd0-3e8b-4df1-84c6-01104522f616" containerID="30ce0fb41d57b8d9015f46071889f7699381f8a114a7a1f73cea8d12533132ab" exitCode=0 Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.423098 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqk2w" event={"ID":"48ca8fd0-3e8b-4df1-84c6-01104522f616","Type":"ContainerDied","Data":"30ce0fb41d57b8d9015f46071889f7699381f8a114a7a1f73cea8d12533132ab"} Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.423133 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqk2w" event={"ID":"48ca8fd0-3e8b-4df1-84c6-01104522f616","Type":"ContainerDied","Data":"d707497fa3ac4be390c2151332b8750c0a6d9c873995ecb6844ee55f6b7a7d62"} Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.423154 4777 scope.go:117] "RemoveContainer" containerID="30ce0fb41d57b8d9015f46071889f7699381f8a114a7a1f73cea8d12533132ab" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.423273 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hqk2w" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.445255 4777 scope.go:117] "RemoveContainer" containerID="e48817a026f61baf17c645029ceb707e359ad6784e666a118679960bdc6eccaf" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.456892 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48ca8fd0-3e8b-4df1-84c6-01104522f616-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.456940 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gg67t\" (UniqueName: \"kubernetes.io/projected/48ca8fd0-3e8b-4df1-84c6-01104522f616-kube-api-access-gg67t\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.462279 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48ca8fd0-3e8b-4df1-84c6-01104522f616-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48ca8fd0-3e8b-4df1-84c6-01104522f616" (UID: "48ca8fd0-3e8b-4df1-84c6-01104522f616"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.476661 4777 scope.go:117] "RemoveContainer" containerID="607f5c5d9f5aec968a92b2544908277da64bba3896e2f515b22fe14d896adfde" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.495224 4777 scope.go:117] "RemoveContainer" containerID="30ce0fb41d57b8d9015f46071889f7699381f8a114a7a1f73cea8d12533132ab" Nov 24 17:04:41 crc kubenswrapper[4777]: E1124 17:04:41.495727 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30ce0fb41d57b8d9015f46071889f7699381f8a114a7a1f73cea8d12533132ab\": container with ID starting with 30ce0fb41d57b8d9015f46071889f7699381f8a114a7a1f73cea8d12533132ab not found: ID does not exist" containerID="30ce0fb41d57b8d9015f46071889f7699381f8a114a7a1f73cea8d12533132ab" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.495772 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30ce0fb41d57b8d9015f46071889f7699381f8a114a7a1f73cea8d12533132ab"} err="failed to get container status \"30ce0fb41d57b8d9015f46071889f7699381f8a114a7a1f73cea8d12533132ab\": rpc error: code = NotFound desc = could not find container \"30ce0fb41d57b8d9015f46071889f7699381f8a114a7a1f73cea8d12533132ab\": container with ID starting with 30ce0fb41d57b8d9015f46071889f7699381f8a114a7a1f73cea8d12533132ab not found: ID does not exist" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.495797 4777 scope.go:117] "RemoveContainer" containerID="e48817a026f61baf17c645029ceb707e359ad6784e666a118679960bdc6eccaf" Nov 24 17:04:41 crc kubenswrapper[4777]: E1124 17:04:41.496164 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e48817a026f61baf17c645029ceb707e359ad6784e666a118679960bdc6eccaf\": container with ID starting with e48817a026f61baf17c645029ceb707e359ad6784e666a118679960bdc6eccaf not found: ID does not exist" containerID="e48817a026f61baf17c645029ceb707e359ad6784e666a118679960bdc6eccaf" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.496217 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e48817a026f61baf17c645029ceb707e359ad6784e666a118679960bdc6eccaf"} err="failed to get container status \"e48817a026f61baf17c645029ceb707e359ad6784e666a118679960bdc6eccaf\": rpc error: code = NotFound desc = could not find container \"e48817a026f61baf17c645029ceb707e359ad6784e666a118679960bdc6eccaf\": container with ID starting with e48817a026f61baf17c645029ceb707e359ad6784e666a118679960bdc6eccaf not found: ID does not exist" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.496250 4777 scope.go:117] "RemoveContainer" containerID="607f5c5d9f5aec968a92b2544908277da64bba3896e2f515b22fe14d896adfde" Nov 24 17:04:41 crc kubenswrapper[4777]: E1124 17:04:41.496555 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"607f5c5d9f5aec968a92b2544908277da64bba3896e2f515b22fe14d896adfde\": container with ID starting with 607f5c5d9f5aec968a92b2544908277da64bba3896e2f515b22fe14d896adfde not found: ID does not exist" containerID="607f5c5d9f5aec968a92b2544908277da64bba3896e2f515b22fe14d896adfde" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.496580 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"607f5c5d9f5aec968a92b2544908277da64bba3896e2f515b22fe14d896adfde"} err="failed to get container status \"607f5c5d9f5aec968a92b2544908277da64bba3896e2f515b22fe14d896adfde\": rpc error: code = NotFound desc = could not find container \"607f5c5d9f5aec968a92b2544908277da64bba3896e2f515b22fe14d896adfde\": container with ID starting with 607f5c5d9f5aec968a92b2544908277da64bba3896e2f515b22fe14d896adfde not found: ID does not exist" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.564082 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48ca8fd0-3e8b-4df1-84c6-01104522f616-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.761585 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hqk2w"] Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.765555 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hqk2w"] Nov 24 17:04:41 crc kubenswrapper[4777]: I1124 17:04:41.783008 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtc8b"] Nov 24 17:04:43 crc kubenswrapper[4777]: I1124 17:04:43.252129 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48ca8fd0-3e8b-4df1-84c6-01104522f616" path="/var/lib/kubelet/pods/48ca8fd0-3e8b-4df1-84c6-01104522f616/volumes" Nov 24 17:04:43 crc kubenswrapper[4777]: I1124 17:04:43.436070 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rtc8b" podUID="549af019-cce7-46bc-a02b-41e1c4e09256" containerName="registry-server" containerID="cri-o://080938aaad95886e016a76524d3b1870ad7e822438b378443454a501f32d80f5" gracePeriod=2 Nov 24 17:04:43 crc kubenswrapper[4777]: I1124 17:04:43.940496 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:04:43 crc kubenswrapper[4777]: I1124 17:04:43.985420 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pszfj"] Nov 24 17:04:43 crc kubenswrapper[4777]: I1124 17:04:43.985824 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pszfj" podUID="0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" containerName="registry-server" containerID="cri-o://0e5419d3f1d8fea39c8aba18c33a9f73e58130a7e761c0a2e348a322013f3b62" gracePeriod=2 Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.107155 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/549af019-cce7-46bc-a02b-41e1c4e09256-catalog-content\") pod \"549af019-cce7-46bc-a02b-41e1c4e09256\" (UID: \"549af019-cce7-46bc-a02b-41e1c4e09256\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.107262 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqpg9\" (UniqueName: \"kubernetes.io/projected/549af019-cce7-46bc-a02b-41e1c4e09256-kube-api-access-cqpg9\") pod \"549af019-cce7-46bc-a02b-41e1c4e09256\" (UID: \"549af019-cce7-46bc-a02b-41e1c4e09256\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.107351 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/549af019-cce7-46bc-a02b-41e1c4e09256-utilities\") pod \"549af019-cce7-46bc-a02b-41e1c4e09256\" (UID: \"549af019-cce7-46bc-a02b-41e1c4e09256\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.109281 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/549af019-cce7-46bc-a02b-41e1c4e09256-utilities" (OuterVolumeSpecName: "utilities") pod "549af019-cce7-46bc-a02b-41e1c4e09256" (UID: "549af019-cce7-46bc-a02b-41e1c4e09256"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.115436 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/549af019-cce7-46bc-a02b-41e1c4e09256-kube-api-access-cqpg9" (OuterVolumeSpecName: "kube-api-access-cqpg9") pod "549af019-cce7-46bc-a02b-41e1c4e09256" (UID: "549af019-cce7-46bc-a02b-41e1c4e09256"). InnerVolumeSpecName "kube-api-access-cqpg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.138863 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/549af019-cce7-46bc-a02b-41e1c4e09256-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "549af019-cce7-46bc-a02b-41e1c4e09256" (UID: "549af019-cce7-46bc-a02b-41e1c4e09256"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.209400 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/549af019-cce7-46bc-a02b-41e1c4e09256-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.209465 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqpg9\" (UniqueName: \"kubernetes.io/projected/549af019-cce7-46bc-a02b-41e1c4e09256-kube-api-access-cqpg9\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.209495 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/549af019-cce7-46bc-a02b-41e1c4e09256-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.360694 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.364379 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" podUID="54a60a6f-27d1-44d8-8372-512191c72f89" containerName="oauth-openshift" containerID="cri-o://234cb8adac9b59e0d6c2b6d4eebda3dd87c18a72f2a002643a0dc3e8742eae0b" gracePeriod=15 Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.442424 4777 generic.go:334] "Generic (PLEG): container finished" podID="549af019-cce7-46bc-a02b-41e1c4e09256" containerID="080938aaad95886e016a76524d3b1870ad7e822438b378443454a501f32d80f5" exitCode=0 Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.442480 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtc8b" event={"ID":"549af019-cce7-46bc-a02b-41e1c4e09256","Type":"ContainerDied","Data":"080938aaad95886e016a76524d3b1870ad7e822438b378443454a501f32d80f5"} Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.442509 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtc8b" event={"ID":"549af019-cce7-46bc-a02b-41e1c4e09256","Type":"ContainerDied","Data":"5112763ee0e4be0cbd46dd34a7d02f87b9120baff9c2b453c2140faea4ad2f3a"} Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.442527 4777 scope.go:117] "RemoveContainer" containerID="080938aaad95886e016a76524d3b1870ad7e822438b378443454a501f32d80f5" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.442635 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtc8b" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.450990 4777 generic.go:334] "Generic (PLEG): container finished" podID="0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" containerID="0e5419d3f1d8fea39c8aba18c33a9f73e58130a7e761c0a2e348a322013f3b62" exitCode=0 Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.451038 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pszfj" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.451048 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pszfj" event={"ID":"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e","Type":"ContainerDied","Data":"0e5419d3f1d8fea39c8aba18c33a9f73e58130a7e761c0a2e348a322013f3b62"} Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.451080 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pszfj" event={"ID":"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e","Type":"ContainerDied","Data":"023a9e8a0aac056e2c519adfab6d6b1442cba0e1912901417e836daf40dbc104"} Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.471677 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtc8b"] Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.471877 4777 scope.go:117] "RemoveContainer" containerID="e6c96623697ef8154219db817e1c8f32fedf0f22a2c9ba97d276601d4b43bfcb" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.475383 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtc8b"] Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.493706 4777 scope.go:117] "RemoveContainer" containerID="bcf14e3d87f861bb6e357707e1c879f17370f22aec6f05f9c9d3c7fd142544b3" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.506313 4777 scope.go:117] "RemoveContainer" containerID="080938aaad95886e016a76524d3b1870ad7e822438b378443454a501f32d80f5" Nov 24 17:04:44 crc kubenswrapper[4777]: E1124 17:04:44.506747 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"080938aaad95886e016a76524d3b1870ad7e822438b378443454a501f32d80f5\": container with ID starting with 080938aaad95886e016a76524d3b1870ad7e822438b378443454a501f32d80f5 not found: ID does not exist" containerID="080938aaad95886e016a76524d3b1870ad7e822438b378443454a501f32d80f5" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.506812 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"080938aaad95886e016a76524d3b1870ad7e822438b378443454a501f32d80f5"} err="failed to get container status \"080938aaad95886e016a76524d3b1870ad7e822438b378443454a501f32d80f5\": rpc error: code = NotFound desc = could not find container \"080938aaad95886e016a76524d3b1870ad7e822438b378443454a501f32d80f5\": container with ID starting with 080938aaad95886e016a76524d3b1870ad7e822438b378443454a501f32d80f5 not found: ID does not exist" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.506841 4777 scope.go:117] "RemoveContainer" containerID="e6c96623697ef8154219db817e1c8f32fedf0f22a2c9ba97d276601d4b43bfcb" Nov 24 17:04:44 crc kubenswrapper[4777]: E1124 17:04:44.507111 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6c96623697ef8154219db817e1c8f32fedf0f22a2c9ba97d276601d4b43bfcb\": container with ID starting with e6c96623697ef8154219db817e1c8f32fedf0f22a2c9ba97d276601d4b43bfcb not found: ID does not exist" containerID="e6c96623697ef8154219db817e1c8f32fedf0f22a2c9ba97d276601d4b43bfcb" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.507135 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6c96623697ef8154219db817e1c8f32fedf0f22a2c9ba97d276601d4b43bfcb"} err="failed to get container status \"e6c96623697ef8154219db817e1c8f32fedf0f22a2c9ba97d276601d4b43bfcb\": rpc error: code = NotFound desc = could not find container \"e6c96623697ef8154219db817e1c8f32fedf0f22a2c9ba97d276601d4b43bfcb\": container with ID starting with e6c96623697ef8154219db817e1c8f32fedf0f22a2c9ba97d276601d4b43bfcb not found: ID does not exist" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.507151 4777 scope.go:117] "RemoveContainer" containerID="bcf14e3d87f861bb6e357707e1c879f17370f22aec6f05f9c9d3c7fd142544b3" Nov 24 17:04:44 crc kubenswrapper[4777]: E1124 17:04:44.507340 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcf14e3d87f861bb6e357707e1c879f17370f22aec6f05f9c9d3c7fd142544b3\": container with ID starting with bcf14e3d87f861bb6e357707e1c879f17370f22aec6f05f9c9d3c7fd142544b3 not found: ID does not exist" containerID="bcf14e3d87f861bb6e357707e1c879f17370f22aec6f05f9c9d3c7fd142544b3" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.507369 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcf14e3d87f861bb6e357707e1c879f17370f22aec6f05f9c9d3c7fd142544b3"} err="failed to get container status \"bcf14e3d87f861bb6e357707e1c879f17370f22aec6f05f9c9d3c7fd142544b3\": rpc error: code = NotFound desc = could not find container \"bcf14e3d87f861bb6e357707e1c879f17370f22aec6f05f9c9d3c7fd142544b3\": container with ID starting with bcf14e3d87f861bb6e357707e1c879f17370f22aec6f05f9c9d3c7fd142544b3 not found: ID does not exist" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.507385 4777 scope.go:117] "RemoveContainer" containerID="0e5419d3f1d8fea39c8aba18c33a9f73e58130a7e761c0a2e348a322013f3b62" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.513473 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6xs7\" (UniqueName: \"kubernetes.io/projected/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-kube-api-access-j6xs7\") pod \"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e\" (UID: \"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.513556 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-catalog-content\") pod \"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e\" (UID: \"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.513641 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-utilities\") pod \"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e\" (UID: \"0a9888ac-755d-4bb9-b5db-07e26a6f1f0e\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.514678 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-utilities" (OuterVolumeSpecName: "utilities") pod "0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" (UID: "0a9888ac-755d-4bb9-b5db-07e26a6f1f0e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.517849 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-kube-api-access-j6xs7" (OuterVolumeSpecName: "kube-api-access-j6xs7") pod "0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" (UID: "0a9888ac-755d-4bb9-b5db-07e26a6f1f0e"). InnerVolumeSpecName "kube-api-access-j6xs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.518659 4777 scope.go:117] "RemoveContainer" containerID="2f726df1ec7b12fd184201d6662815d8deeea50748906599a1f5f7c9962bd77d" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.543276 4777 scope.go:117] "RemoveContainer" containerID="7c3619857d5fbabe9964bb5a543c3959ab6d3dc74e19a10c0d8169668abc71e9" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.558510 4777 scope.go:117] "RemoveContainer" containerID="0e5419d3f1d8fea39c8aba18c33a9f73e58130a7e761c0a2e348a322013f3b62" Nov 24 17:04:44 crc kubenswrapper[4777]: E1124 17:04:44.559043 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e5419d3f1d8fea39c8aba18c33a9f73e58130a7e761c0a2e348a322013f3b62\": container with ID starting with 0e5419d3f1d8fea39c8aba18c33a9f73e58130a7e761c0a2e348a322013f3b62 not found: ID does not exist" containerID="0e5419d3f1d8fea39c8aba18c33a9f73e58130a7e761c0a2e348a322013f3b62" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.559087 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e5419d3f1d8fea39c8aba18c33a9f73e58130a7e761c0a2e348a322013f3b62"} err="failed to get container status \"0e5419d3f1d8fea39c8aba18c33a9f73e58130a7e761c0a2e348a322013f3b62\": rpc error: code = NotFound desc = could not find container \"0e5419d3f1d8fea39c8aba18c33a9f73e58130a7e761c0a2e348a322013f3b62\": container with ID starting with 0e5419d3f1d8fea39c8aba18c33a9f73e58130a7e761c0a2e348a322013f3b62 not found: ID does not exist" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.559107 4777 scope.go:117] "RemoveContainer" containerID="2f726df1ec7b12fd184201d6662815d8deeea50748906599a1f5f7c9962bd77d" Nov 24 17:04:44 crc kubenswrapper[4777]: E1124 17:04:44.559412 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f726df1ec7b12fd184201d6662815d8deeea50748906599a1f5f7c9962bd77d\": container with ID starting with 2f726df1ec7b12fd184201d6662815d8deeea50748906599a1f5f7c9962bd77d not found: ID does not exist" containerID="2f726df1ec7b12fd184201d6662815d8deeea50748906599a1f5f7c9962bd77d" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.559458 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f726df1ec7b12fd184201d6662815d8deeea50748906599a1f5f7c9962bd77d"} err="failed to get container status \"2f726df1ec7b12fd184201d6662815d8deeea50748906599a1f5f7c9962bd77d\": rpc error: code = NotFound desc = could not find container \"2f726df1ec7b12fd184201d6662815d8deeea50748906599a1f5f7c9962bd77d\": container with ID starting with 2f726df1ec7b12fd184201d6662815d8deeea50748906599a1f5f7c9962bd77d not found: ID does not exist" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.559494 4777 scope.go:117] "RemoveContainer" containerID="7c3619857d5fbabe9964bb5a543c3959ab6d3dc74e19a10c0d8169668abc71e9" Nov 24 17:04:44 crc kubenswrapper[4777]: E1124 17:04:44.559742 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c3619857d5fbabe9964bb5a543c3959ab6d3dc74e19a10c0d8169668abc71e9\": container with ID starting with 7c3619857d5fbabe9964bb5a543c3959ab6d3dc74e19a10c0d8169668abc71e9 not found: ID does not exist" containerID="7c3619857d5fbabe9964bb5a543c3959ab6d3dc74e19a10c0d8169668abc71e9" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.559771 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c3619857d5fbabe9964bb5a543c3959ab6d3dc74e19a10c0d8169668abc71e9"} err="failed to get container status \"7c3619857d5fbabe9964bb5a543c3959ab6d3dc74e19a10c0d8169668abc71e9\": rpc error: code = NotFound desc = could not find container \"7c3619857d5fbabe9964bb5a543c3959ab6d3dc74e19a10c0d8169668abc71e9\": container with ID starting with 7c3619857d5fbabe9964bb5a543c3959ab6d3dc74e19a10c0d8169668abc71e9 not found: ID does not exist" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.615657 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6xs7\" (UniqueName: \"kubernetes.io/projected/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-kube-api-access-j6xs7\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.615685 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.620209 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" (UID: "0a9888ac-755d-4bb9-b5db-07e26a6f1f0e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.706336 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.719977 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.791940 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pszfj"] Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.794881 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pszfj"] Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.821083 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-idp-0-file-data\") pod \"54a60a6f-27d1-44d8-8372-512191c72f89\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.821160 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-session\") pod \"54a60a6f-27d1-44d8-8372-512191c72f89\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.821189 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sqqw\" (UniqueName: \"kubernetes.io/projected/54a60a6f-27d1-44d8-8372-512191c72f89-kube-api-access-8sqqw\") pod \"54a60a6f-27d1-44d8-8372-512191c72f89\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.821235 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-error\") pod \"54a60a6f-27d1-44d8-8372-512191c72f89\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.821261 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-ocp-branding-template\") pod \"54a60a6f-27d1-44d8-8372-512191c72f89\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.821287 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-login\") pod \"54a60a6f-27d1-44d8-8372-512191c72f89\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.821306 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-provider-selection\") pod \"54a60a6f-27d1-44d8-8372-512191c72f89\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.821341 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-trusted-ca-bundle\") pod \"54a60a6f-27d1-44d8-8372-512191c72f89\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.821367 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-router-certs\") pod \"54a60a6f-27d1-44d8-8372-512191c72f89\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.821394 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-cliconfig\") pod \"54a60a6f-27d1-44d8-8372-512191c72f89\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.821418 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-serving-cert\") pod \"54a60a6f-27d1-44d8-8372-512191c72f89\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.821447 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-service-ca\") pod \"54a60a6f-27d1-44d8-8372-512191c72f89\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.821475 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54a60a6f-27d1-44d8-8372-512191c72f89-audit-dir\") pod \"54a60a6f-27d1-44d8-8372-512191c72f89\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.821513 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-audit-policies\") pod \"54a60a6f-27d1-44d8-8372-512191c72f89\" (UID: \"54a60a6f-27d1-44d8-8372-512191c72f89\") " Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.822435 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "54a60a6f-27d1-44d8-8372-512191c72f89" (UID: "54a60a6f-27d1-44d8-8372-512191c72f89"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.824912 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "54a60a6f-27d1-44d8-8372-512191c72f89" (UID: "54a60a6f-27d1-44d8-8372-512191c72f89"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.825408 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "54a60a6f-27d1-44d8-8372-512191c72f89" (UID: "54a60a6f-27d1-44d8-8372-512191c72f89"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.825745 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/54a60a6f-27d1-44d8-8372-512191c72f89-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "54a60a6f-27d1-44d8-8372-512191c72f89" (UID: "54a60a6f-27d1-44d8-8372-512191c72f89"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.826672 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "54a60a6f-27d1-44d8-8372-512191c72f89" (UID: "54a60a6f-27d1-44d8-8372-512191c72f89"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.829557 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "54a60a6f-27d1-44d8-8372-512191c72f89" (UID: "54a60a6f-27d1-44d8-8372-512191c72f89"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.830417 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "54a60a6f-27d1-44d8-8372-512191c72f89" (UID: "54a60a6f-27d1-44d8-8372-512191c72f89"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.830914 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54a60a6f-27d1-44d8-8372-512191c72f89-kube-api-access-8sqqw" (OuterVolumeSpecName: "kube-api-access-8sqqw") pod "54a60a6f-27d1-44d8-8372-512191c72f89" (UID: "54a60a6f-27d1-44d8-8372-512191c72f89"). InnerVolumeSpecName "kube-api-access-8sqqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.831611 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "54a60a6f-27d1-44d8-8372-512191c72f89" (UID: "54a60a6f-27d1-44d8-8372-512191c72f89"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.831894 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "54a60a6f-27d1-44d8-8372-512191c72f89" (UID: "54a60a6f-27d1-44d8-8372-512191c72f89"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.832590 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "54a60a6f-27d1-44d8-8372-512191c72f89" (UID: "54a60a6f-27d1-44d8-8372-512191c72f89"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.832825 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "54a60a6f-27d1-44d8-8372-512191c72f89" (UID: "54a60a6f-27d1-44d8-8372-512191c72f89"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.833257 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "54a60a6f-27d1-44d8-8372-512191c72f89" (UID: "54a60a6f-27d1-44d8-8372-512191c72f89"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.832680 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "54a60a6f-27d1-44d8-8372-512191c72f89" (UID: "54a60a6f-27d1-44d8-8372-512191c72f89"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.922774 4777 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54a60a6f-27d1-44d8-8372-512191c72f89-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.922809 4777 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.922818 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.922828 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.922838 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sqqw\" (UniqueName: \"kubernetes.io/projected/54a60a6f-27d1-44d8-8372-512191c72f89-kube-api-access-8sqqw\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.922847 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.922856 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.922866 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.922875 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.922884 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.922893 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.922901 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.922911 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:44 crc kubenswrapper[4777]: I1124 17:04:44.922919 4777 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/54a60a6f-27d1-44d8-8372-512191c72f89-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.255128 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" path="/var/lib/kubelet/pods/0a9888ac-755d-4bb9-b5db-07e26a6f1f0e/volumes" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.255947 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="549af019-cce7-46bc-a02b-41e1c4e09256" path="/var/lib/kubelet/pods/549af019-cce7-46bc-a02b-41e1c4e09256/volumes" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.461089 4777 generic.go:334] "Generic (PLEG): container finished" podID="54a60a6f-27d1-44d8-8372-512191c72f89" containerID="234cb8adac9b59e0d6c2b6d4eebda3dd87c18a72f2a002643a0dc3e8742eae0b" exitCode=0 Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.461201 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" event={"ID":"54a60a6f-27d1-44d8-8372-512191c72f89","Type":"ContainerDied","Data":"234cb8adac9b59e0d6c2b6d4eebda3dd87c18a72f2a002643a0dc3e8742eae0b"} Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.461229 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.461279 4777 scope.go:117] "RemoveContainer" containerID="234cb8adac9b59e0d6c2b6d4eebda3dd87c18a72f2a002643a0dc3e8742eae0b" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.461267 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7hwq6" event={"ID":"54a60a6f-27d1-44d8-8372-512191c72f89","Type":"ContainerDied","Data":"6ab0e58e616b76783bdbdb740d6a80a7ed5c41de305fa1e108f581c3c53349ea"} Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.480777 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7hwq6"] Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.490070 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7hwq6"] Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.494002 4777 scope.go:117] "RemoveContainer" containerID="234cb8adac9b59e0d6c2b6d4eebda3dd87c18a72f2a002643a0dc3e8742eae0b" Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.494675 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"234cb8adac9b59e0d6c2b6d4eebda3dd87c18a72f2a002643a0dc3e8742eae0b\": container with ID starting with 234cb8adac9b59e0d6c2b6d4eebda3dd87c18a72f2a002643a0dc3e8742eae0b not found: ID does not exist" containerID="234cb8adac9b59e0d6c2b6d4eebda3dd87c18a72f2a002643a0dc3e8742eae0b" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.494728 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"234cb8adac9b59e0d6c2b6d4eebda3dd87c18a72f2a002643a0dc3e8742eae0b"} err="failed to get container status \"234cb8adac9b59e0d6c2b6d4eebda3dd87c18a72f2a002643a0dc3e8742eae0b\": rpc error: code = NotFound desc = could not find container \"234cb8adac9b59e0d6c2b6d4eebda3dd87c18a72f2a002643a0dc3e8742eae0b\": container with ID starting with 234cb8adac9b59e0d6c2b6d4eebda3dd87c18a72f2a002643a0dc3e8742eae0b not found: ID does not exist" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.607454 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5844cf768-v59r8"] Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.608430 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" containerName="extract-utilities" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.608467 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" containerName="extract-utilities" Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.608497 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48ca8fd0-3e8b-4df1-84c6-01104522f616" containerName="extract-content" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.608514 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="48ca8fd0-3e8b-4df1-84c6-01104522f616" containerName="extract-content" Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.608537 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eebaa807-fbd5-40d7-ac22-02523cbc4fc2" containerName="extract-content" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.608554 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="eebaa807-fbd5-40d7-ac22-02523cbc4fc2" containerName="extract-content" Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.608581 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0493374-1b14-4216-bf96-b81a619212e5" containerName="pruner" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.608595 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0493374-1b14-4216-bf96-b81a619212e5" containerName="pruner" Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.608615 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="404fb408-cca3-49ee-acb8-69f336a96d0a" containerName="pruner" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.608629 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="404fb408-cca3-49ee-acb8-69f336a96d0a" containerName="pruner" Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.608646 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a60a6f-27d1-44d8-8372-512191c72f89" containerName="oauth-openshift" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.608665 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a60a6f-27d1-44d8-8372-512191c72f89" containerName="oauth-openshift" Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.608683 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" containerName="extract-content" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.608699 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" containerName="extract-content" Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.608725 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" containerName="registry-server" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.608741 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" containerName="registry-server" Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.608774 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="549af019-cce7-46bc-a02b-41e1c4e09256" containerName="registry-server" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.608790 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="549af019-cce7-46bc-a02b-41e1c4e09256" containerName="registry-server" Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.608807 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48ca8fd0-3e8b-4df1-84c6-01104522f616" containerName="registry-server" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.608840 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="48ca8fd0-3e8b-4df1-84c6-01104522f616" containerName="registry-server" Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.608866 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eebaa807-fbd5-40d7-ac22-02523cbc4fc2" containerName="extract-utilities" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.608881 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="eebaa807-fbd5-40d7-ac22-02523cbc4fc2" containerName="extract-utilities" Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.608905 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="549af019-cce7-46bc-a02b-41e1c4e09256" containerName="extract-utilities" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.608923 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="549af019-cce7-46bc-a02b-41e1c4e09256" containerName="extract-utilities" Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.608993 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eebaa807-fbd5-40d7-ac22-02523cbc4fc2" containerName="registry-server" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.609013 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="eebaa807-fbd5-40d7-ac22-02523cbc4fc2" containerName="registry-server" Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.609035 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="549af019-cce7-46bc-a02b-41e1c4e09256" containerName="extract-content" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.609051 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="549af019-cce7-46bc-a02b-41e1c4e09256" containerName="extract-content" Nov 24 17:04:45 crc kubenswrapper[4777]: E1124 17:04:45.609076 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48ca8fd0-3e8b-4df1-84c6-01104522f616" containerName="extract-utilities" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.609095 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="48ca8fd0-3e8b-4df1-84c6-01104522f616" containerName="extract-utilities" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.609358 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="eebaa807-fbd5-40d7-ac22-02523cbc4fc2" containerName="registry-server" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.609389 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="48ca8fd0-3e8b-4df1-84c6-01104522f616" containerName="registry-server" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.609445 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a9888ac-755d-4bb9-b5db-07e26a6f1f0e" containerName="registry-server" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.609474 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0493374-1b14-4216-bf96-b81a619212e5" containerName="pruner" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.609497 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="54a60a6f-27d1-44d8-8372-512191c72f89" containerName="oauth-openshift" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.609520 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="549af019-cce7-46bc-a02b-41e1c4e09256" containerName="registry-server" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.609546 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="404fb408-cca3-49ee-acb8-69f336a96d0a" containerName="pruner" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.610280 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.614110 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.614511 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.624431 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.624826 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.625943 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.626116 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.626239 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.626331 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.626426 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.627442 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.628320 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.628475 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.636858 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5844cf768-v59r8"] Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.639732 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.642818 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.653917 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.734374 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.734433 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.734667 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.734741 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-audit-dir\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.734804 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-user-template-error\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.734871 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-session\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.735043 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-service-ca\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.735110 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-router-certs\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.735161 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.735207 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-user-template-login\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.735332 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.735399 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsqb5\" (UniqueName: \"kubernetes.io/projected/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-kube-api-access-fsqb5\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.735442 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.735487 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-audit-policies\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.837354 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-service-ca\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.837441 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-router-certs\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.837488 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.837544 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-user-template-login\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.837635 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.837709 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsqb5\" (UniqueName: \"kubernetes.io/projected/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-kube-api-access-fsqb5\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.837758 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.837822 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-audit-policies\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.837945 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.838044 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.838143 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.838196 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-audit-dir\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.838242 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-user-template-error\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.838290 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-session\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.838553 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-service-ca\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.838965 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-audit-dir\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.840167 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-audit-policies\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.841445 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.841451 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.844926 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.845958 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-user-template-login\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.846412 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-session\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.847060 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-router-certs\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.847210 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.847662 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.848392 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-user-template-error\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.850610 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.866074 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsqb5\" (UniqueName: \"kubernetes.io/projected/5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa-kube-api-access-fsqb5\") pod \"oauth-openshift-5844cf768-v59r8\" (UID: \"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa\") " pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:45 crc kubenswrapper[4777]: I1124 17:04:45.952555 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:46 crc kubenswrapper[4777]: I1124 17:04:46.204079 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5844cf768-v59r8"] Nov 24 17:04:46 crc kubenswrapper[4777]: I1124 17:04:46.469576 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" event={"ID":"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa","Type":"ContainerStarted","Data":"2ff4ddba3b769c75acee59bdd0e361886c93e3d58e736d096a3e3d22c3acd090"} Nov 24 17:04:47 crc kubenswrapper[4777]: I1124 17:04:47.252131 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54a60a6f-27d1-44d8-8372-512191c72f89" path="/var/lib/kubelet/pods/54a60a6f-27d1-44d8-8372-512191c72f89/volumes" Nov 24 17:04:47 crc kubenswrapper[4777]: I1124 17:04:47.476481 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" event={"ID":"5d7b7ffa-41bb-45f1-9b6e-d74ec04ff6aa","Type":"ContainerStarted","Data":"47c06d4f0647754900efcd206f279d27f6a79006d3bfab1e8eba5c4c4015b99f"} Nov 24 17:04:47 crc kubenswrapper[4777]: I1124 17:04:47.476731 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:47 crc kubenswrapper[4777]: I1124 17:04:47.483396 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" Nov 24 17:04:47 crc kubenswrapper[4777]: I1124 17:04:47.498040 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5844cf768-v59r8" podStartSLOduration=28.498018354 podStartE2EDuration="28.498018354s" podCreationTimestamp="2025-11-24 17:04:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:04:47.49492246 +0000 UTC m=+245.654117509" watchObservedRunningTime="2025-11-24 17:04:47.498018354 +0000 UTC m=+245.657213413" Nov 24 17:05:07 crc kubenswrapper[4777]: I1124 17:05:07.934178 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-shz64"] Nov 24 17:05:07 crc kubenswrapper[4777]: I1124 17:05:07.934818 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-shz64" podUID="b93bcc12-1487-4c0f-9da2-fab9c61df5c9" containerName="registry-server" containerID="cri-o://bbd16d4ddad153f2a9d67eb28915908ece9b183567ba4018777fed0cacd79014" gracePeriod=30 Nov 24 17:05:07 crc kubenswrapper[4777]: I1124 17:05:07.948161 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tr7gm"] Nov 24 17:05:07 crc kubenswrapper[4777]: I1124 17:05:07.949235 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tr7gm" podUID="6203d413-faf0-4476-9a91-345ba4f1772f" containerName="registry-server" containerID="cri-o://ecc95bdd713cd10e0c52925e7fedd4aec58c072eb482b738a1e2c5c81300be65" gracePeriod=30 Nov 24 17:05:07 crc kubenswrapper[4777]: I1124 17:05:07.957149 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6tcwx"] Nov 24 17:05:07 crc kubenswrapper[4777]: I1124 17:05:07.957323 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" podUID="e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab" containerName="marketplace-operator" containerID="cri-o://e6d9e2fde864cd94bc88afcc48e3729605ae1e54de574f289c93c7dff087f737" gracePeriod=30 Nov 24 17:05:07 crc kubenswrapper[4777]: I1124 17:05:07.971127 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mbg7"] Nov 24 17:05:07 crc kubenswrapper[4777]: I1124 17:05:07.971495 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9mbg7" podUID="caacdae7-8c80-4fcb-85de-78dcc1cecf59" containerName="registry-server" containerID="cri-o://9e115b1137cd5fa9f36d9d39f4ac62d1cf331b53a68ed9932ac520442b78f452" gracePeriod=30 Nov 24 17:05:07 crc kubenswrapper[4777]: I1124 17:05:07.979693 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-26x9k"] Nov 24 17:05:07 crc kubenswrapper[4777]: I1124 17:05:07.979980 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-26x9k" podUID="3af02e22-079e-4f81-9f7f-c5b55591387e" containerName="registry-server" containerID="cri-o://a8a8ef411766bb3846abc92b8692f7065a3f509759ebdaf136f9fbac914a9d1d" gracePeriod=30 Nov 24 17:05:07 crc kubenswrapper[4777]: I1124 17:05:07.989768 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5dsxp"] Nov 24 17:05:07 crc kubenswrapper[4777]: I1124 17:05:07.990597 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" Nov 24 17:05:07 crc kubenswrapper[4777]: I1124 17:05:07.995387 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5dsxp"] Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.145021 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5dsxp\" (UID: \"3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da\") " pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.145066 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cblzm\" (UniqueName: \"kubernetes.io/projected/3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da-kube-api-access-cblzm\") pod \"marketplace-operator-79b997595-5dsxp\" (UID: \"3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da\") " pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.145094 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5dsxp\" (UID: \"3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da\") " pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.246856 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cblzm\" (UniqueName: \"kubernetes.io/projected/3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da-kube-api-access-cblzm\") pod \"marketplace-operator-79b997595-5dsxp\" (UID: \"3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da\") " pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.246901 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5dsxp\" (UID: \"3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da\") " pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.246946 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5dsxp\" (UID: \"3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da\") " pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.249079 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5dsxp\" (UID: \"3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da\") " pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.261093 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5dsxp\" (UID: \"3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da\") " pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.261675 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cblzm\" (UniqueName: \"kubernetes.io/projected/3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da-kube-api-access-cblzm\") pod \"marketplace-operator-79b997595-5dsxp\" (UID: \"3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da\") " pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.385414 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.388662 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.398666 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.400932 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.412145 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.439824 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.550614 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjlwm\" (UniqueName: \"kubernetes.io/projected/3af02e22-079e-4f81-9f7f-c5b55591387e-kube-api-access-jjlwm\") pod \"3af02e22-079e-4f81-9f7f-c5b55591387e\" (UID: \"3af02e22-079e-4f81-9f7f-c5b55591387e\") " Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.550951 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-marketplace-trusted-ca\") pod \"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab\" (UID: \"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab\") " Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.551003 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-catalog-content\") pod \"b93bcc12-1487-4c0f-9da2-fab9c61df5c9\" (UID: \"b93bcc12-1487-4c0f-9da2-fab9c61df5c9\") " Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.551037 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6qnd\" (UniqueName: \"kubernetes.io/projected/6203d413-faf0-4476-9a91-345ba4f1772f-kube-api-access-b6qnd\") pod \"6203d413-faf0-4476-9a91-345ba4f1772f\" (UID: \"6203d413-faf0-4476-9a91-345ba4f1772f\") " Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.551065 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caacdae7-8c80-4fcb-85de-78dcc1cecf59-catalog-content\") pod \"caacdae7-8c80-4fcb-85de-78dcc1cecf59\" (UID: \"caacdae7-8c80-4fcb-85de-78dcc1cecf59\") " Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.551081 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af02e22-079e-4f81-9f7f-c5b55591387e-catalog-content\") pod \"3af02e22-079e-4f81-9f7f-c5b55591387e\" (UID: \"3af02e22-079e-4f81-9f7f-c5b55591387e\") " Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.551114 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caacdae7-8c80-4fcb-85de-78dcc1cecf59-utilities\") pod \"caacdae7-8c80-4fcb-85de-78dcc1cecf59\" (UID: \"caacdae7-8c80-4fcb-85de-78dcc1cecf59\") " Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.551133 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af02e22-079e-4f81-9f7f-c5b55591387e-utilities\") pod \"3af02e22-079e-4f81-9f7f-c5b55591387e\" (UID: \"3af02e22-079e-4f81-9f7f-c5b55591387e\") " Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.551156 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-utilities\") pod \"b93bcc12-1487-4c0f-9da2-fab9c61df5c9\" (UID: \"b93bcc12-1487-4c0f-9da2-fab9c61df5c9\") " Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.551180 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmh6j\" (UniqueName: \"kubernetes.io/projected/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-kube-api-access-zmh6j\") pod \"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab\" (UID: \"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab\") " Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.551205 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbflk\" (UniqueName: \"kubernetes.io/projected/caacdae7-8c80-4fcb-85de-78dcc1cecf59-kube-api-access-gbflk\") pod \"caacdae7-8c80-4fcb-85de-78dcc1cecf59\" (UID: \"caacdae7-8c80-4fcb-85de-78dcc1cecf59\") " Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.551244 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6203d413-faf0-4476-9a91-345ba4f1772f-catalog-content\") pod \"6203d413-faf0-4476-9a91-345ba4f1772f\" (UID: \"6203d413-faf0-4476-9a91-345ba4f1772f\") " Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.551280 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6203d413-faf0-4476-9a91-345ba4f1772f-utilities\") pod \"6203d413-faf0-4476-9a91-345ba4f1772f\" (UID: \"6203d413-faf0-4476-9a91-345ba4f1772f\") " Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.551302 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-marketplace-operator-metrics\") pod \"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab\" (UID: \"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab\") " Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.551323 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8ng5\" (UniqueName: \"kubernetes.io/projected/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-kube-api-access-w8ng5\") pod \"b93bcc12-1487-4c0f-9da2-fab9c61df5c9\" (UID: \"b93bcc12-1487-4c0f-9da2-fab9c61df5c9\") " Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.552197 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caacdae7-8c80-4fcb-85de-78dcc1cecf59-utilities" (OuterVolumeSpecName: "utilities") pod "caacdae7-8c80-4fcb-85de-78dcc1cecf59" (UID: "caacdae7-8c80-4fcb-85de-78dcc1cecf59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.552402 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab" (UID: "e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.552667 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3af02e22-079e-4f81-9f7f-c5b55591387e-utilities" (OuterVolumeSpecName: "utilities") pod "3af02e22-079e-4f81-9f7f-c5b55591387e" (UID: "3af02e22-079e-4f81-9f7f-c5b55591387e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.552696 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6203d413-faf0-4476-9a91-345ba4f1772f-utilities" (OuterVolumeSpecName: "utilities") pod "6203d413-faf0-4476-9a91-345ba4f1772f" (UID: "6203d413-faf0-4476-9a91-345ba4f1772f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.554190 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-utilities" (OuterVolumeSpecName: "utilities") pod "b93bcc12-1487-4c0f-9da2-fab9c61df5c9" (UID: "b93bcc12-1487-4c0f-9da2-fab9c61df5c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.563393 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3af02e22-079e-4f81-9f7f-c5b55591387e-kube-api-access-jjlwm" (OuterVolumeSpecName: "kube-api-access-jjlwm") pod "3af02e22-079e-4f81-9f7f-c5b55591387e" (UID: "3af02e22-079e-4f81-9f7f-c5b55591387e"). InnerVolumeSpecName "kube-api-access-jjlwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.563791 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6203d413-faf0-4476-9a91-345ba4f1772f-kube-api-access-b6qnd" (OuterVolumeSpecName: "kube-api-access-b6qnd") pod "6203d413-faf0-4476-9a91-345ba4f1772f" (UID: "6203d413-faf0-4476-9a91-345ba4f1772f"). InnerVolumeSpecName "kube-api-access-b6qnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.564880 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-kube-api-access-w8ng5" (OuterVolumeSpecName: "kube-api-access-w8ng5") pod "b93bcc12-1487-4c0f-9da2-fab9c61df5c9" (UID: "b93bcc12-1487-4c0f-9da2-fab9c61df5c9"). InnerVolumeSpecName "kube-api-access-w8ng5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.565046 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caacdae7-8c80-4fcb-85de-78dcc1cecf59-kube-api-access-gbflk" (OuterVolumeSpecName: "kube-api-access-gbflk") pod "caacdae7-8c80-4fcb-85de-78dcc1cecf59" (UID: "caacdae7-8c80-4fcb-85de-78dcc1cecf59"). InnerVolumeSpecName "kube-api-access-gbflk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.565355 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab" (UID: "e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.574898 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-kube-api-access-zmh6j" (OuterVolumeSpecName: "kube-api-access-zmh6j") pod "e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab" (UID: "e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab"). InnerVolumeSpecName "kube-api-access-zmh6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.600818 4777 generic.go:334] "Generic (PLEG): container finished" podID="e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab" containerID="e6d9e2fde864cd94bc88afcc48e3729605ae1e54de574f289c93c7dff087f737" exitCode=0 Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.600870 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" event={"ID":"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab","Type":"ContainerDied","Data":"e6d9e2fde864cd94bc88afcc48e3729605ae1e54de574f289c93c7dff087f737"} Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.600895 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" event={"ID":"e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab","Type":"ContainerDied","Data":"486b6adbef148a2d440eb69531d54e8c3cbfb18fb236123ccf99497bd5a95b33"} Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.600909 4777 scope.go:117] "RemoveContainer" containerID="e6d9e2fde864cd94bc88afcc48e3729605ae1e54de574f289c93c7dff087f737" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.601018 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6tcwx" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.604275 4777 generic.go:334] "Generic (PLEG): container finished" podID="6203d413-faf0-4476-9a91-345ba4f1772f" containerID="ecc95bdd713cd10e0c52925e7fedd4aec58c072eb482b738a1e2c5c81300be65" exitCode=0 Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.604341 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr7gm" event={"ID":"6203d413-faf0-4476-9a91-345ba4f1772f","Type":"ContainerDied","Data":"ecc95bdd713cd10e0c52925e7fedd4aec58c072eb482b738a1e2c5c81300be65"} Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.604362 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr7gm" event={"ID":"6203d413-faf0-4476-9a91-345ba4f1772f","Type":"ContainerDied","Data":"7ff90c859ffd8ba86b369c0ac90a167b40bbd5de37eb9f298fc766383d228b4b"} Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.604431 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tr7gm" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.610311 4777 generic.go:334] "Generic (PLEG): container finished" podID="b93bcc12-1487-4c0f-9da2-fab9c61df5c9" containerID="bbd16d4ddad153f2a9d67eb28915908ece9b183567ba4018777fed0cacd79014" exitCode=0 Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.610387 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-shz64" event={"ID":"b93bcc12-1487-4c0f-9da2-fab9c61df5c9","Type":"ContainerDied","Data":"bbd16d4ddad153f2a9d67eb28915908ece9b183567ba4018777fed0cacd79014"} Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.610408 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-shz64" event={"ID":"b93bcc12-1487-4c0f-9da2-fab9c61df5c9","Type":"ContainerDied","Data":"c45cfe50b77cf43e9df4b5a9c22a188d037546d16dc6c03ab73633547571c1fc"} Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.610461 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-shz64" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.617223 4777 generic.go:334] "Generic (PLEG): container finished" podID="caacdae7-8c80-4fcb-85de-78dcc1cecf59" containerID="9e115b1137cd5fa9f36d9d39f4ac62d1cf331b53a68ed9932ac520442b78f452" exitCode=0 Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.617277 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mbg7" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.617279 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mbg7" event={"ID":"caacdae7-8c80-4fcb-85de-78dcc1cecf59","Type":"ContainerDied","Data":"9e115b1137cd5fa9f36d9d39f4ac62d1cf331b53a68ed9932ac520442b78f452"} Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.617362 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mbg7" event={"ID":"caacdae7-8c80-4fcb-85de-78dcc1cecf59","Type":"ContainerDied","Data":"ee65df9a8c3d0007610acee378b398fee47c9218d74cb910edfcf11a7dfa776b"} Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.621506 4777 generic.go:334] "Generic (PLEG): container finished" podID="3af02e22-079e-4f81-9f7f-c5b55591387e" containerID="a8a8ef411766bb3846abc92b8692f7065a3f509759ebdaf136f9fbac914a9d1d" exitCode=0 Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.621536 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-26x9k" event={"ID":"3af02e22-079e-4f81-9f7f-c5b55591387e","Type":"ContainerDied","Data":"a8a8ef411766bb3846abc92b8692f7065a3f509759ebdaf136f9fbac914a9d1d"} Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.621585 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-26x9k" event={"ID":"3af02e22-079e-4f81-9f7f-c5b55591387e","Type":"ContainerDied","Data":"7be700959cf6fffed9fe6040d85d381bfdbae7639808b45a40d22209b8152789"} Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.621675 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-26x9k" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.624022 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caacdae7-8c80-4fcb-85de-78dcc1cecf59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "caacdae7-8c80-4fcb-85de-78dcc1cecf59" (UID: "caacdae7-8c80-4fcb-85de-78dcc1cecf59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.625541 4777 scope.go:117] "RemoveContainer" containerID="e6d9e2fde864cd94bc88afcc48e3729605ae1e54de574f289c93c7dff087f737" Nov 24 17:05:08 crc kubenswrapper[4777]: E1124 17:05:08.627276 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6d9e2fde864cd94bc88afcc48e3729605ae1e54de574f289c93c7dff087f737\": container with ID starting with e6d9e2fde864cd94bc88afcc48e3729605ae1e54de574f289c93c7dff087f737 not found: ID does not exist" containerID="e6d9e2fde864cd94bc88afcc48e3729605ae1e54de574f289c93c7dff087f737" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.627367 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d9e2fde864cd94bc88afcc48e3729605ae1e54de574f289c93c7dff087f737"} err="failed to get container status \"e6d9e2fde864cd94bc88afcc48e3729605ae1e54de574f289c93c7dff087f737\": rpc error: code = NotFound desc = could not find container \"e6d9e2fde864cd94bc88afcc48e3729605ae1e54de574f289c93c7dff087f737\": container with ID starting with e6d9e2fde864cd94bc88afcc48e3729605ae1e54de574f289c93c7dff087f737 not found: ID does not exist" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.627396 4777 scope.go:117] "RemoveContainer" containerID="ecc95bdd713cd10e0c52925e7fedd4aec58c072eb482b738a1e2c5c81300be65" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.652786 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjlwm\" (UniqueName: \"kubernetes.io/projected/3af02e22-079e-4f81-9f7f-c5b55591387e-kube-api-access-jjlwm\") on node \"crc\" DevicePath \"\"" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.652871 4777 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.652882 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6qnd\" (UniqueName: \"kubernetes.io/projected/6203d413-faf0-4476-9a91-345ba4f1772f-kube-api-access-b6qnd\") on node \"crc\" DevicePath \"\"" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.652909 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caacdae7-8c80-4fcb-85de-78dcc1cecf59-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.652939 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caacdae7-8c80-4fcb-85de-78dcc1cecf59-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.652949 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af02e22-079e-4f81-9f7f-c5b55591387e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.652980 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.652999 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmh6j\" (UniqueName: \"kubernetes.io/projected/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-kube-api-access-zmh6j\") on node \"crc\" DevicePath \"\"" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.653008 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbflk\" (UniqueName: \"kubernetes.io/projected/caacdae7-8c80-4fcb-85de-78dcc1cecf59-kube-api-access-gbflk\") on node \"crc\" DevicePath \"\"" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.653018 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6203d413-faf0-4476-9a91-345ba4f1772f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.653038 4777 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.653048 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8ng5\" (UniqueName: \"kubernetes.io/projected/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-kube-api-access-w8ng5\") on node \"crc\" DevicePath \"\"" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.655303 4777 scope.go:117] "RemoveContainer" containerID="322b76ed6de4ddbafd27e4249776fe3ab148c238d62681752038a73991ff9f9c" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.667114 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b93bcc12-1487-4c0f-9da2-fab9c61df5c9" (UID: "b93bcc12-1487-4c0f-9da2-fab9c61df5c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.671032 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6tcwx"] Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.678820 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6tcwx"] Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.691328 4777 scope.go:117] "RemoveContainer" containerID="3c97b6302bf5a6f62b361f7f5a5c20b7bcb13a1b40d6502b4c78a62332bfdbeb" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.692127 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3af02e22-079e-4f81-9f7f-c5b55591387e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3af02e22-079e-4f81-9f7f-c5b55591387e" (UID: "3af02e22-079e-4f81-9f7f-c5b55591387e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.699563 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5dsxp"] Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.700031 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6203d413-faf0-4476-9a91-345ba4f1772f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6203d413-faf0-4476-9a91-345ba4f1772f" (UID: "6203d413-faf0-4476-9a91-345ba4f1772f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.705320 4777 scope.go:117] "RemoveContainer" containerID="ecc95bdd713cd10e0c52925e7fedd4aec58c072eb482b738a1e2c5c81300be65" Nov 24 17:05:08 crc kubenswrapper[4777]: E1124 17:05:08.705790 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecc95bdd713cd10e0c52925e7fedd4aec58c072eb482b738a1e2c5c81300be65\": container with ID starting with ecc95bdd713cd10e0c52925e7fedd4aec58c072eb482b738a1e2c5c81300be65 not found: ID does not exist" containerID="ecc95bdd713cd10e0c52925e7fedd4aec58c072eb482b738a1e2c5c81300be65" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.705825 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecc95bdd713cd10e0c52925e7fedd4aec58c072eb482b738a1e2c5c81300be65"} err="failed to get container status \"ecc95bdd713cd10e0c52925e7fedd4aec58c072eb482b738a1e2c5c81300be65\": rpc error: code = NotFound desc = could not find container \"ecc95bdd713cd10e0c52925e7fedd4aec58c072eb482b738a1e2c5c81300be65\": container with ID starting with ecc95bdd713cd10e0c52925e7fedd4aec58c072eb482b738a1e2c5c81300be65 not found: ID does not exist" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.705854 4777 scope.go:117] "RemoveContainer" containerID="322b76ed6de4ddbafd27e4249776fe3ab148c238d62681752038a73991ff9f9c" Nov 24 17:05:08 crc kubenswrapper[4777]: E1124 17:05:08.706112 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"322b76ed6de4ddbafd27e4249776fe3ab148c238d62681752038a73991ff9f9c\": container with ID starting with 322b76ed6de4ddbafd27e4249776fe3ab148c238d62681752038a73991ff9f9c not found: ID does not exist" containerID="322b76ed6de4ddbafd27e4249776fe3ab148c238d62681752038a73991ff9f9c" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.706131 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"322b76ed6de4ddbafd27e4249776fe3ab148c238d62681752038a73991ff9f9c"} err="failed to get container status \"322b76ed6de4ddbafd27e4249776fe3ab148c238d62681752038a73991ff9f9c\": rpc error: code = NotFound desc = could not find container \"322b76ed6de4ddbafd27e4249776fe3ab148c238d62681752038a73991ff9f9c\": container with ID starting with 322b76ed6de4ddbafd27e4249776fe3ab148c238d62681752038a73991ff9f9c not found: ID does not exist" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.706145 4777 scope.go:117] "RemoveContainer" containerID="3c97b6302bf5a6f62b361f7f5a5c20b7bcb13a1b40d6502b4c78a62332bfdbeb" Nov 24 17:05:08 crc kubenswrapper[4777]: E1124 17:05:08.706379 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c97b6302bf5a6f62b361f7f5a5c20b7bcb13a1b40d6502b4c78a62332bfdbeb\": container with ID starting with 3c97b6302bf5a6f62b361f7f5a5c20b7bcb13a1b40d6502b4c78a62332bfdbeb not found: ID does not exist" containerID="3c97b6302bf5a6f62b361f7f5a5c20b7bcb13a1b40d6502b4c78a62332bfdbeb" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.706399 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c97b6302bf5a6f62b361f7f5a5c20b7bcb13a1b40d6502b4c78a62332bfdbeb"} err="failed to get container status \"3c97b6302bf5a6f62b361f7f5a5c20b7bcb13a1b40d6502b4c78a62332bfdbeb\": rpc error: code = NotFound desc = could not find container \"3c97b6302bf5a6f62b361f7f5a5c20b7bcb13a1b40d6502b4c78a62332bfdbeb\": container with ID starting with 3c97b6302bf5a6f62b361f7f5a5c20b7bcb13a1b40d6502b4c78a62332bfdbeb not found: ID does not exist" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.706412 4777 scope.go:117] "RemoveContainer" containerID="bbd16d4ddad153f2a9d67eb28915908ece9b183567ba4018777fed0cacd79014" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.718670 4777 scope.go:117] "RemoveContainer" containerID="77413ac0483f9186775c6ac946883282de8436766b13e3fe00a3906d741ddd07" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.750316 4777 scope.go:117] "RemoveContainer" containerID="2bf66f64adac1f0f975225dd4b6f5be282bdbe508c9dbb3987a9633417089d23" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.754576 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b93bcc12-1487-4c0f-9da2-fab9c61df5c9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.754598 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af02e22-079e-4f81-9f7f-c5b55591387e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.754608 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6203d413-faf0-4476-9a91-345ba4f1772f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.765602 4777 scope.go:117] "RemoveContainer" containerID="bbd16d4ddad153f2a9d67eb28915908ece9b183567ba4018777fed0cacd79014" Nov 24 17:05:08 crc kubenswrapper[4777]: E1124 17:05:08.766075 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbd16d4ddad153f2a9d67eb28915908ece9b183567ba4018777fed0cacd79014\": container with ID starting with bbd16d4ddad153f2a9d67eb28915908ece9b183567ba4018777fed0cacd79014 not found: ID does not exist" containerID="bbd16d4ddad153f2a9d67eb28915908ece9b183567ba4018777fed0cacd79014" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.766120 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbd16d4ddad153f2a9d67eb28915908ece9b183567ba4018777fed0cacd79014"} err="failed to get container status \"bbd16d4ddad153f2a9d67eb28915908ece9b183567ba4018777fed0cacd79014\": rpc error: code = NotFound desc = could not find container \"bbd16d4ddad153f2a9d67eb28915908ece9b183567ba4018777fed0cacd79014\": container with ID starting with bbd16d4ddad153f2a9d67eb28915908ece9b183567ba4018777fed0cacd79014 not found: ID does not exist" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.766153 4777 scope.go:117] "RemoveContainer" containerID="77413ac0483f9186775c6ac946883282de8436766b13e3fe00a3906d741ddd07" Nov 24 17:05:08 crc kubenswrapper[4777]: E1124 17:05:08.766560 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77413ac0483f9186775c6ac946883282de8436766b13e3fe00a3906d741ddd07\": container with ID starting with 77413ac0483f9186775c6ac946883282de8436766b13e3fe00a3906d741ddd07 not found: ID does not exist" containerID="77413ac0483f9186775c6ac946883282de8436766b13e3fe00a3906d741ddd07" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.766601 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77413ac0483f9186775c6ac946883282de8436766b13e3fe00a3906d741ddd07"} err="failed to get container status \"77413ac0483f9186775c6ac946883282de8436766b13e3fe00a3906d741ddd07\": rpc error: code = NotFound desc = could not find container \"77413ac0483f9186775c6ac946883282de8436766b13e3fe00a3906d741ddd07\": container with ID starting with 77413ac0483f9186775c6ac946883282de8436766b13e3fe00a3906d741ddd07 not found: ID does not exist" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.766629 4777 scope.go:117] "RemoveContainer" containerID="2bf66f64adac1f0f975225dd4b6f5be282bdbe508c9dbb3987a9633417089d23" Nov 24 17:05:08 crc kubenswrapper[4777]: E1124 17:05:08.767052 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bf66f64adac1f0f975225dd4b6f5be282bdbe508c9dbb3987a9633417089d23\": container with ID starting with 2bf66f64adac1f0f975225dd4b6f5be282bdbe508c9dbb3987a9633417089d23 not found: ID does not exist" containerID="2bf66f64adac1f0f975225dd4b6f5be282bdbe508c9dbb3987a9633417089d23" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.767088 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bf66f64adac1f0f975225dd4b6f5be282bdbe508c9dbb3987a9633417089d23"} err="failed to get container status \"2bf66f64adac1f0f975225dd4b6f5be282bdbe508c9dbb3987a9633417089d23\": rpc error: code = NotFound desc = could not find container \"2bf66f64adac1f0f975225dd4b6f5be282bdbe508c9dbb3987a9633417089d23\": container with ID starting with 2bf66f64adac1f0f975225dd4b6f5be282bdbe508c9dbb3987a9633417089d23 not found: ID does not exist" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.767105 4777 scope.go:117] "RemoveContainer" containerID="9e115b1137cd5fa9f36d9d39f4ac62d1cf331b53a68ed9932ac520442b78f452" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.789151 4777 scope.go:117] "RemoveContainer" containerID="45f4b0a7abe4dc38fd70eea11f2e3b2240cc21d3987197911ec0eaac43c49574" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.802232 4777 scope.go:117] "RemoveContainer" containerID="104c59ff8e9c8c45c0d58cc125ef0a8cfbc8f98dffb354be7bde4b7afc82c20c" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.816123 4777 scope.go:117] "RemoveContainer" containerID="9e115b1137cd5fa9f36d9d39f4ac62d1cf331b53a68ed9932ac520442b78f452" Nov 24 17:05:08 crc kubenswrapper[4777]: E1124 17:05:08.816569 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e115b1137cd5fa9f36d9d39f4ac62d1cf331b53a68ed9932ac520442b78f452\": container with ID starting with 9e115b1137cd5fa9f36d9d39f4ac62d1cf331b53a68ed9932ac520442b78f452 not found: ID does not exist" containerID="9e115b1137cd5fa9f36d9d39f4ac62d1cf331b53a68ed9932ac520442b78f452" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.816614 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e115b1137cd5fa9f36d9d39f4ac62d1cf331b53a68ed9932ac520442b78f452"} err="failed to get container status \"9e115b1137cd5fa9f36d9d39f4ac62d1cf331b53a68ed9932ac520442b78f452\": rpc error: code = NotFound desc = could not find container \"9e115b1137cd5fa9f36d9d39f4ac62d1cf331b53a68ed9932ac520442b78f452\": container with ID starting with 9e115b1137cd5fa9f36d9d39f4ac62d1cf331b53a68ed9932ac520442b78f452 not found: ID does not exist" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.816649 4777 scope.go:117] "RemoveContainer" containerID="45f4b0a7abe4dc38fd70eea11f2e3b2240cc21d3987197911ec0eaac43c49574" Nov 24 17:05:08 crc kubenswrapper[4777]: E1124 17:05:08.816945 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45f4b0a7abe4dc38fd70eea11f2e3b2240cc21d3987197911ec0eaac43c49574\": container with ID starting with 45f4b0a7abe4dc38fd70eea11f2e3b2240cc21d3987197911ec0eaac43c49574 not found: ID does not exist" containerID="45f4b0a7abe4dc38fd70eea11f2e3b2240cc21d3987197911ec0eaac43c49574" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.817024 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45f4b0a7abe4dc38fd70eea11f2e3b2240cc21d3987197911ec0eaac43c49574"} err="failed to get container status \"45f4b0a7abe4dc38fd70eea11f2e3b2240cc21d3987197911ec0eaac43c49574\": rpc error: code = NotFound desc = could not find container \"45f4b0a7abe4dc38fd70eea11f2e3b2240cc21d3987197911ec0eaac43c49574\": container with ID starting with 45f4b0a7abe4dc38fd70eea11f2e3b2240cc21d3987197911ec0eaac43c49574 not found: ID does not exist" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.817040 4777 scope.go:117] "RemoveContainer" containerID="104c59ff8e9c8c45c0d58cc125ef0a8cfbc8f98dffb354be7bde4b7afc82c20c" Nov 24 17:05:08 crc kubenswrapper[4777]: E1124 17:05:08.817288 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"104c59ff8e9c8c45c0d58cc125ef0a8cfbc8f98dffb354be7bde4b7afc82c20c\": container with ID starting with 104c59ff8e9c8c45c0d58cc125ef0a8cfbc8f98dffb354be7bde4b7afc82c20c not found: ID does not exist" containerID="104c59ff8e9c8c45c0d58cc125ef0a8cfbc8f98dffb354be7bde4b7afc82c20c" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.817315 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"104c59ff8e9c8c45c0d58cc125ef0a8cfbc8f98dffb354be7bde4b7afc82c20c"} err="failed to get container status \"104c59ff8e9c8c45c0d58cc125ef0a8cfbc8f98dffb354be7bde4b7afc82c20c\": rpc error: code = NotFound desc = could not find container \"104c59ff8e9c8c45c0d58cc125ef0a8cfbc8f98dffb354be7bde4b7afc82c20c\": container with ID starting with 104c59ff8e9c8c45c0d58cc125ef0a8cfbc8f98dffb354be7bde4b7afc82c20c not found: ID does not exist" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.817331 4777 scope.go:117] "RemoveContainer" containerID="a8a8ef411766bb3846abc92b8692f7065a3f509759ebdaf136f9fbac914a9d1d" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.832554 4777 scope.go:117] "RemoveContainer" containerID="f16a0f8f2ee0d857a2639480c01badf387ba3ad1d6521fd9db5d38a82d0e3606" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.848010 4777 scope.go:117] "RemoveContainer" containerID="5ee96844507a69d62274a029e8bd7bcdb4a4e4a695de59442ef99b7b1dcd9f73" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.859493 4777 scope.go:117] "RemoveContainer" containerID="a8a8ef411766bb3846abc92b8692f7065a3f509759ebdaf136f9fbac914a9d1d" Nov 24 17:05:08 crc kubenswrapper[4777]: E1124 17:05:08.859882 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8a8ef411766bb3846abc92b8692f7065a3f509759ebdaf136f9fbac914a9d1d\": container with ID starting with a8a8ef411766bb3846abc92b8692f7065a3f509759ebdaf136f9fbac914a9d1d not found: ID does not exist" containerID="a8a8ef411766bb3846abc92b8692f7065a3f509759ebdaf136f9fbac914a9d1d" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.859925 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8a8ef411766bb3846abc92b8692f7065a3f509759ebdaf136f9fbac914a9d1d"} err="failed to get container status \"a8a8ef411766bb3846abc92b8692f7065a3f509759ebdaf136f9fbac914a9d1d\": rpc error: code = NotFound desc = could not find container \"a8a8ef411766bb3846abc92b8692f7065a3f509759ebdaf136f9fbac914a9d1d\": container with ID starting with a8a8ef411766bb3846abc92b8692f7065a3f509759ebdaf136f9fbac914a9d1d not found: ID does not exist" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.859960 4777 scope.go:117] "RemoveContainer" containerID="f16a0f8f2ee0d857a2639480c01badf387ba3ad1d6521fd9db5d38a82d0e3606" Nov 24 17:05:08 crc kubenswrapper[4777]: E1124 17:05:08.860298 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f16a0f8f2ee0d857a2639480c01badf387ba3ad1d6521fd9db5d38a82d0e3606\": container with ID starting with f16a0f8f2ee0d857a2639480c01badf387ba3ad1d6521fd9db5d38a82d0e3606 not found: ID does not exist" containerID="f16a0f8f2ee0d857a2639480c01badf387ba3ad1d6521fd9db5d38a82d0e3606" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.860332 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f16a0f8f2ee0d857a2639480c01badf387ba3ad1d6521fd9db5d38a82d0e3606"} err="failed to get container status \"f16a0f8f2ee0d857a2639480c01badf387ba3ad1d6521fd9db5d38a82d0e3606\": rpc error: code = NotFound desc = could not find container \"f16a0f8f2ee0d857a2639480c01badf387ba3ad1d6521fd9db5d38a82d0e3606\": container with ID starting with f16a0f8f2ee0d857a2639480c01badf387ba3ad1d6521fd9db5d38a82d0e3606 not found: ID does not exist" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.860356 4777 scope.go:117] "RemoveContainer" containerID="5ee96844507a69d62274a029e8bd7bcdb4a4e4a695de59442ef99b7b1dcd9f73" Nov 24 17:05:08 crc kubenswrapper[4777]: E1124 17:05:08.860599 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ee96844507a69d62274a029e8bd7bcdb4a4e4a695de59442ef99b7b1dcd9f73\": container with ID starting with 5ee96844507a69d62274a029e8bd7bcdb4a4e4a695de59442ef99b7b1dcd9f73 not found: ID does not exist" containerID="5ee96844507a69d62274a029e8bd7bcdb4a4e4a695de59442ef99b7b1dcd9f73" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.860621 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ee96844507a69d62274a029e8bd7bcdb4a4e4a695de59442ef99b7b1dcd9f73"} err="failed to get container status \"5ee96844507a69d62274a029e8bd7bcdb4a4e4a695de59442ef99b7b1dcd9f73\": rpc error: code = NotFound desc = could not find container \"5ee96844507a69d62274a029e8bd7bcdb4a4e4a695de59442ef99b7b1dcd9f73\": container with ID starting with 5ee96844507a69d62274a029e8bd7bcdb4a4e4a695de59442ef99b7b1dcd9f73 not found: ID does not exist" Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.936333 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tr7gm"] Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.940429 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tr7gm"] Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.949564 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-shz64"] Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.956016 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-shz64"] Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.964407 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-26x9k"] Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.967897 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-26x9k"] Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.972947 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mbg7"] Nov 24 17:05:08 crc kubenswrapper[4777]: I1124 17:05:08.977134 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mbg7"] Nov 24 17:05:09 crc kubenswrapper[4777]: I1124 17:05:09.251287 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3af02e22-079e-4f81-9f7f-c5b55591387e" path="/var/lib/kubelet/pods/3af02e22-079e-4f81-9f7f-c5b55591387e/volumes" Nov 24 17:05:09 crc kubenswrapper[4777]: I1124 17:05:09.252204 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6203d413-faf0-4476-9a91-345ba4f1772f" path="/var/lib/kubelet/pods/6203d413-faf0-4476-9a91-345ba4f1772f/volumes" Nov 24 17:05:09 crc kubenswrapper[4777]: I1124 17:05:09.252744 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b93bcc12-1487-4c0f-9da2-fab9c61df5c9" path="/var/lib/kubelet/pods/b93bcc12-1487-4c0f-9da2-fab9c61df5c9/volumes" Nov 24 17:05:09 crc kubenswrapper[4777]: I1124 17:05:09.253689 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caacdae7-8c80-4fcb-85de-78dcc1cecf59" path="/var/lib/kubelet/pods/caacdae7-8c80-4fcb-85de-78dcc1cecf59/volumes" Nov 24 17:05:09 crc kubenswrapper[4777]: I1124 17:05:09.254281 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab" path="/var/lib/kubelet/pods/e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab/volumes" Nov 24 17:05:09 crc kubenswrapper[4777]: I1124 17:05:09.632977 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" event={"ID":"3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da","Type":"ContainerStarted","Data":"4124edf8397ecdbad3ac97dd3d92ad0915133d191f0610a48f4163bfa89e3f98"} Nov 24 17:05:09 crc kubenswrapper[4777]: I1124 17:05:09.633017 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" event={"ID":"3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da","Type":"ContainerStarted","Data":"ebb5bc3042db5824be31f3aa3cd429186d28af453727c49afc3bdf4ada1aae36"} Nov 24 17:05:09 crc kubenswrapper[4777]: I1124 17:05:09.633522 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" Nov 24 17:05:09 crc kubenswrapper[4777]: I1124 17:05:09.637860 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" Nov 24 17:05:09 crc kubenswrapper[4777]: I1124 17:05:09.668716 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5dsxp" podStartSLOduration=2.668698361 podStartE2EDuration="2.668698361s" podCreationTimestamp="2025-11-24 17:05:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:05:09.648080576 +0000 UTC m=+267.807275625" watchObservedRunningTime="2025-11-24 17:05:09.668698361 +0000 UTC m=+267.827893420" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149103 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hwftb"] Nov 24 17:05:10 crc kubenswrapper[4777]: E1124 17:05:10.149329 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b93bcc12-1487-4c0f-9da2-fab9c61df5c9" containerName="extract-utilities" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149344 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="b93bcc12-1487-4c0f-9da2-fab9c61df5c9" containerName="extract-utilities" Nov 24 17:05:10 crc kubenswrapper[4777]: E1124 17:05:10.149356 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3af02e22-079e-4f81-9f7f-c5b55591387e" containerName="extract-content" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149361 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="3af02e22-079e-4f81-9f7f-c5b55591387e" containerName="extract-content" Nov 24 17:05:10 crc kubenswrapper[4777]: E1124 17:05:10.149371 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caacdae7-8c80-4fcb-85de-78dcc1cecf59" containerName="extract-utilities" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149377 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="caacdae7-8c80-4fcb-85de-78dcc1cecf59" containerName="extract-utilities" Nov 24 17:05:10 crc kubenswrapper[4777]: E1124 17:05:10.149386 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6203d413-faf0-4476-9a91-345ba4f1772f" containerName="registry-server" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149392 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="6203d413-faf0-4476-9a91-345ba4f1772f" containerName="registry-server" Nov 24 17:05:10 crc kubenswrapper[4777]: E1124 17:05:10.149400 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caacdae7-8c80-4fcb-85de-78dcc1cecf59" containerName="registry-server" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149406 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="caacdae7-8c80-4fcb-85de-78dcc1cecf59" containerName="registry-server" Nov 24 17:05:10 crc kubenswrapper[4777]: E1124 17:05:10.149414 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6203d413-faf0-4476-9a91-345ba4f1772f" containerName="extract-content" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149420 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="6203d413-faf0-4476-9a91-345ba4f1772f" containerName="extract-content" Nov 24 17:05:10 crc kubenswrapper[4777]: E1124 17:05:10.149432 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3af02e22-079e-4f81-9f7f-c5b55591387e" containerName="registry-server" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149437 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="3af02e22-079e-4f81-9f7f-c5b55591387e" containerName="registry-server" Nov 24 17:05:10 crc kubenswrapper[4777]: E1124 17:05:10.149445 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab" containerName="marketplace-operator" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149451 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab" containerName="marketplace-operator" Nov 24 17:05:10 crc kubenswrapper[4777]: E1124 17:05:10.149461 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caacdae7-8c80-4fcb-85de-78dcc1cecf59" containerName="extract-content" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149470 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="caacdae7-8c80-4fcb-85de-78dcc1cecf59" containerName="extract-content" Nov 24 17:05:10 crc kubenswrapper[4777]: E1124 17:05:10.149487 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6203d413-faf0-4476-9a91-345ba4f1772f" containerName="extract-utilities" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149494 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="6203d413-faf0-4476-9a91-345ba4f1772f" containerName="extract-utilities" Nov 24 17:05:10 crc kubenswrapper[4777]: E1124 17:05:10.149502 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3af02e22-079e-4f81-9f7f-c5b55591387e" containerName="extract-utilities" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149510 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="3af02e22-079e-4f81-9f7f-c5b55591387e" containerName="extract-utilities" Nov 24 17:05:10 crc kubenswrapper[4777]: E1124 17:05:10.149517 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b93bcc12-1487-4c0f-9da2-fab9c61df5c9" containerName="registry-server" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149525 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="b93bcc12-1487-4c0f-9da2-fab9c61df5c9" containerName="registry-server" Nov 24 17:05:10 crc kubenswrapper[4777]: E1124 17:05:10.149534 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b93bcc12-1487-4c0f-9da2-fab9c61df5c9" containerName="extract-content" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149541 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="b93bcc12-1487-4c0f-9da2-fab9c61df5c9" containerName="extract-content" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149659 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="3af02e22-079e-4f81-9f7f-c5b55591387e" containerName="registry-server" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149669 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="6203d413-faf0-4476-9a91-345ba4f1772f" containerName="registry-server" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149678 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5da5de3-01bc-4e9c-a0e5-8dc3a34cdaab" containerName="marketplace-operator" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149685 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="caacdae7-8c80-4fcb-85de-78dcc1cecf59" containerName="registry-server" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.149694 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="b93bcc12-1487-4c0f-9da2-fab9c61df5c9" containerName="registry-server" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.150339 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hwftb" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.154280 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.168260 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hwftb"] Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.180222 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27733c5c-8536-4f12-8aa0-e5a9d16e8f88-catalog-content\") pod \"redhat-marketplace-hwftb\" (UID: \"27733c5c-8536-4f12-8aa0-e5a9d16e8f88\") " pod="openshift-marketplace/redhat-marketplace-hwftb" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.180293 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-895fg\" (UniqueName: \"kubernetes.io/projected/27733c5c-8536-4f12-8aa0-e5a9d16e8f88-kube-api-access-895fg\") pod \"redhat-marketplace-hwftb\" (UID: \"27733c5c-8536-4f12-8aa0-e5a9d16e8f88\") " pod="openshift-marketplace/redhat-marketplace-hwftb" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.180343 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27733c5c-8536-4f12-8aa0-e5a9d16e8f88-utilities\") pod \"redhat-marketplace-hwftb\" (UID: \"27733c5c-8536-4f12-8aa0-e5a9d16e8f88\") " pod="openshift-marketplace/redhat-marketplace-hwftb" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.281912 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27733c5c-8536-4f12-8aa0-e5a9d16e8f88-catalog-content\") pod \"redhat-marketplace-hwftb\" (UID: \"27733c5c-8536-4f12-8aa0-e5a9d16e8f88\") " pod="openshift-marketplace/redhat-marketplace-hwftb" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.281996 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-895fg\" (UniqueName: \"kubernetes.io/projected/27733c5c-8536-4f12-8aa0-e5a9d16e8f88-kube-api-access-895fg\") pod \"redhat-marketplace-hwftb\" (UID: \"27733c5c-8536-4f12-8aa0-e5a9d16e8f88\") " pod="openshift-marketplace/redhat-marketplace-hwftb" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.282065 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27733c5c-8536-4f12-8aa0-e5a9d16e8f88-utilities\") pod \"redhat-marketplace-hwftb\" (UID: \"27733c5c-8536-4f12-8aa0-e5a9d16e8f88\") " pod="openshift-marketplace/redhat-marketplace-hwftb" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.282434 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27733c5c-8536-4f12-8aa0-e5a9d16e8f88-catalog-content\") pod \"redhat-marketplace-hwftb\" (UID: \"27733c5c-8536-4f12-8aa0-e5a9d16e8f88\") " pod="openshift-marketplace/redhat-marketplace-hwftb" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.282549 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27733c5c-8536-4f12-8aa0-e5a9d16e8f88-utilities\") pod \"redhat-marketplace-hwftb\" (UID: \"27733c5c-8536-4f12-8aa0-e5a9d16e8f88\") " pod="openshift-marketplace/redhat-marketplace-hwftb" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.311410 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-895fg\" (UniqueName: \"kubernetes.io/projected/27733c5c-8536-4f12-8aa0-e5a9d16e8f88-kube-api-access-895fg\") pod \"redhat-marketplace-hwftb\" (UID: \"27733c5c-8536-4f12-8aa0-e5a9d16e8f88\") " pod="openshift-marketplace/redhat-marketplace-hwftb" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.353424 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g7f5l"] Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.354417 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g7f5l" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.356242 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.361520 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g7f5l"] Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.383524 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1af51c54-3a40-406e-8743-32b0c04d018f-utilities\") pod \"redhat-operators-g7f5l\" (UID: \"1af51c54-3a40-406e-8743-32b0c04d018f\") " pod="openshift-marketplace/redhat-operators-g7f5l" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.383740 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmv8d\" (UniqueName: \"kubernetes.io/projected/1af51c54-3a40-406e-8743-32b0c04d018f-kube-api-access-dmv8d\") pod \"redhat-operators-g7f5l\" (UID: \"1af51c54-3a40-406e-8743-32b0c04d018f\") " pod="openshift-marketplace/redhat-operators-g7f5l" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.383856 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1af51c54-3a40-406e-8743-32b0c04d018f-catalog-content\") pod \"redhat-operators-g7f5l\" (UID: \"1af51c54-3a40-406e-8743-32b0c04d018f\") " pod="openshift-marketplace/redhat-operators-g7f5l" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.470175 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hwftb" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.484788 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1af51c54-3a40-406e-8743-32b0c04d018f-catalog-content\") pod \"redhat-operators-g7f5l\" (UID: \"1af51c54-3a40-406e-8743-32b0c04d018f\") " pod="openshift-marketplace/redhat-operators-g7f5l" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.484856 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1af51c54-3a40-406e-8743-32b0c04d018f-utilities\") pod \"redhat-operators-g7f5l\" (UID: \"1af51c54-3a40-406e-8743-32b0c04d018f\") " pod="openshift-marketplace/redhat-operators-g7f5l" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.484939 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmv8d\" (UniqueName: \"kubernetes.io/projected/1af51c54-3a40-406e-8743-32b0c04d018f-kube-api-access-dmv8d\") pod \"redhat-operators-g7f5l\" (UID: \"1af51c54-3a40-406e-8743-32b0c04d018f\") " pod="openshift-marketplace/redhat-operators-g7f5l" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.485442 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1af51c54-3a40-406e-8743-32b0c04d018f-utilities\") pod \"redhat-operators-g7f5l\" (UID: \"1af51c54-3a40-406e-8743-32b0c04d018f\") " pod="openshift-marketplace/redhat-operators-g7f5l" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.486125 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1af51c54-3a40-406e-8743-32b0c04d018f-catalog-content\") pod \"redhat-operators-g7f5l\" (UID: \"1af51c54-3a40-406e-8743-32b0c04d018f\") " pod="openshift-marketplace/redhat-operators-g7f5l" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.505368 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmv8d\" (UniqueName: \"kubernetes.io/projected/1af51c54-3a40-406e-8743-32b0c04d018f-kube-api-access-dmv8d\") pod \"redhat-operators-g7f5l\" (UID: \"1af51c54-3a40-406e-8743-32b0c04d018f\") " pod="openshift-marketplace/redhat-operators-g7f5l" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.678571 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g7f5l" Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.868342 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hwftb"] Nov 24 17:05:10 crc kubenswrapper[4777]: W1124 17:05:10.874274 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27733c5c_8536_4f12_8aa0_e5a9d16e8f88.slice/crio-4f5aa62fa6eec7b703f731463af0420dae6d0b45ba6127beb10467a8cf7d6dfd WatchSource:0}: Error finding container 4f5aa62fa6eec7b703f731463af0420dae6d0b45ba6127beb10467a8cf7d6dfd: Status 404 returned error can't find the container with id 4f5aa62fa6eec7b703f731463af0420dae6d0b45ba6127beb10467a8cf7d6dfd Nov 24 17:05:10 crc kubenswrapper[4777]: I1124 17:05:10.879885 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g7f5l"] Nov 24 17:05:11 crc kubenswrapper[4777]: I1124 17:05:11.648322 4777 generic.go:334] "Generic (PLEG): container finished" podID="1af51c54-3a40-406e-8743-32b0c04d018f" containerID="a93997d213c1f5303f379df32d90e61c6a2fb143404e5ac9eba24997e13319f9" exitCode=0 Nov 24 17:05:11 crc kubenswrapper[4777]: I1124 17:05:11.648424 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7f5l" event={"ID":"1af51c54-3a40-406e-8743-32b0c04d018f","Type":"ContainerDied","Data":"a93997d213c1f5303f379df32d90e61c6a2fb143404e5ac9eba24997e13319f9"} Nov 24 17:05:11 crc kubenswrapper[4777]: I1124 17:05:11.648629 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7f5l" event={"ID":"1af51c54-3a40-406e-8743-32b0c04d018f","Type":"ContainerStarted","Data":"57db8eb61b8e2976d5c1e65207a9554dd9d895ca72af44d8153b2ff2ab00e536"} Nov 24 17:05:11 crc kubenswrapper[4777]: I1124 17:05:11.652362 4777 generic.go:334] "Generic (PLEG): container finished" podID="27733c5c-8536-4f12-8aa0-e5a9d16e8f88" containerID="b8d2469fd0f8bf3cb578cf824f95a794ff2a82e77b20e21d0ea94992224c2ac8" exitCode=0 Nov 24 17:05:11 crc kubenswrapper[4777]: I1124 17:05:11.652439 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hwftb" event={"ID":"27733c5c-8536-4f12-8aa0-e5a9d16e8f88","Type":"ContainerDied","Data":"b8d2469fd0f8bf3cb578cf824f95a794ff2a82e77b20e21d0ea94992224c2ac8"} Nov 24 17:05:11 crc kubenswrapper[4777]: I1124 17:05:11.652493 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hwftb" event={"ID":"27733c5c-8536-4f12-8aa0-e5a9d16e8f88","Type":"ContainerStarted","Data":"4f5aa62fa6eec7b703f731463af0420dae6d0b45ba6127beb10467a8cf7d6dfd"} Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.552397 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5m4fr"] Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.556000 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5m4fr"] Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.556272 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5m4fr" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.558712 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.621614 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aa2fa6f-bd34-4154-b666-455e7fea71ae-utilities\") pod \"certified-operators-5m4fr\" (UID: \"6aa2fa6f-bd34-4154-b666-455e7fea71ae\") " pod="openshift-marketplace/certified-operators-5m4fr" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.621761 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brzwg\" (UniqueName: \"kubernetes.io/projected/6aa2fa6f-bd34-4154-b666-455e7fea71ae-kube-api-access-brzwg\") pod \"certified-operators-5m4fr\" (UID: \"6aa2fa6f-bd34-4154-b666-455e7fea71ae\") " pod="openshift-marketplace/certified-operators-5m4fr" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.621858 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aa2fa6f-bd34-4154-b666-455e7fea71ae-catalog-content\") pod \"certified-operators-5m4fr\" (UID: \"6aa2fa6f-bd34-4154-b666-455e7fea71ae\") " pod="openshift-marketplace/certified-operators-5m4fr" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.666766 4777 generic.go:334] "Generic (PLEG): container finished" podID="27733c5c-8536-4f12-8aa0-e5a9d16e8f88" containerID="85784b1500f936bb8f2de0f96e09e31e38d2094110d8673afa548d833100f0e4" exitCode=0 Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.666805 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hwftb" event={"ID":"27733c5c-8536-4f12-8aa0-e5a9d16e8f88","Type":"ContainerDied","Data":"85784b1500f936bb8f2de0f96e09e31e38d2094110d8673afa548d833100f0e4"} Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.723604 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aa2fa6f-bd34-4154-b666-455e7fea71ae-catalog-content\") pod \"certified-operators-5m4fr\" (UID: \"6aa2fa6f-bd34-4154-b666-455e7fea71ae\") " pod="openshift-marketplace/certified-operators-5m4fr" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.724070 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aa2fa6f-bd34-4154-b666-455e7fea71ae-catalog-content\") pod \"certified-operators-5m4fr\" (UID: \"6aa2fa6f-bd34-4154-b666-455e7fea71ae\") " pod="openshift-marketplace/certified-operators-5m4fr" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.724287 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aa2fa6f-bd34-4154-b666-455e7fea71ae-utilities\") pod \"certified-operators-5m4fr\" (UID: \"6aa2fa6f-bd34-4154-b666-455e7fea71ae\") " pod="openshift-marketplace/certified-operators-5m4fr" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.724397 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brzwg\" (UniqueName: \"kubernetes.io/projected/6aa2fa6f-bd34-4154-b666-455e7fea71ae-kube-api-access-brzwg\") pod \"certified-operators-5m4fr\" (UID: \"6aa2fa6f-bd34-4154-b666-455e7fea71ae\") " pod="openshift-marketplace/certified-operators-5m4fr" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.725183 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aa2fa6f-bd34-4154-b666-455e7fea71ae-utilities\") pod \"certified-operators-5m4fr\" (UID: \"6aa2fa6f-bd34-4154-b666-455e7fea71ae\") " pod="openshift-marketplace/certified-operators-5m4fr" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.742922 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brzwg\" (UniqueName: \"kubernetes.io/projected/6aa2fa6f-bd34-4154-b666-455e7fea71ae-kube-api-access-brzwg\") pod \"certified-operators-5m4fr\" (UID: \"6aa2fa6f-bd34-4154-b666-455e7fea71ae\") " pod="openshift-marketplace/certified-operators-5m4fr" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.751047 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h54gs"] Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.752191 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h54gs" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.755292 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.761629 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h54gs"] Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.825494 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddnsc\" (UniqueName: \"kubernetes.io/projected/30623966-678a-4124-bc99-02d4ccfacb85-kube-api-access-ddnsc\") pod \"community-operators-h54gs\" (UID: \"30623966-678a-4124-bc99-02d4ccfacb85\") " pod="openshift-marketplace/community-operators-h54gs" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.825575 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30623966-678a-4124-bc99-02d4ccfacb85-utilities\") pod \"community-operators-h54gs\" (UID: \"30623966-678a-4124-bc99-02d4ccfacb85\") " pod="openshift-marketplace/community-operators-h54gs" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.825616 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30623966-678a-4124-bc99-02d4ccfacb85-catalog-content\") pod \"community-operators-h54gs\" (UID: \"30623966-678a-4124-bc99-02d4ccfacb85\") " pod="openshift-marketplace/community-operators-h54gs" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.928816 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5m4fr" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.929366 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30623966-678a-4124-bc99-02d4ccfacb85-catalog-content\") pod \"community-operators-h54gs\" (UID: \"30623966-678a-4124-bc99-02d4ccfacb85\") " pod="openshift-marketplace/community-operators-h54gs" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.929456 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddnsc\" (UniqueName: \"kubernetes.io/projected/30623966-678a-4124-bc99-02d4ccfacb85-kube-api-access-ddnsc\") pod \"community-operators-h54gs\" (UID: \"30623966-678a-4124-bc99-02d4ccfacb85\") " pod="openshift-marketplace/community-operators-h54gs" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.929529 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30623966-678a-4124-bc99-02d4ccfacb85-utilities\") pod \"community-operators-h54gs\" (UID: \"30623966-678a-4124-bc99-02d4ccfacb85\") " pod="openshift-marketplace/community-operators-h54gs" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.929875 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30623966-678a-4124-bc99-02d4ccfacb85-catalog-content\") pod \"community-operators-h54gs\" (UID: \"30623966-678a-4124-bc99-02d4ccfacb85\") " pod="openshift-marketplace/community-operators-h54gs" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.929997 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30623966-678a-4124-bc99-02d4ccfacb85-utilities\") pod \"community-operators-h54gs\" (UID: \"30623966-678a-4124-bc99-02d4ccfacb85\") " pod="openshift-marketplace/community-operators-h54gs" Nov 24 17:05:12 crc kubenswrapper[4777]: I1124 17:05:12.944743 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddnsc\" (UniqueName: \"kubernetes.io/projected/30623966-678a-4124-bc99-02d4ccfacb85-kube-api-access-ddnsc\") pod \"community-operators-h54gs\" (UID: \"30623966-678a-4124-bc99-02d4ccfacb85\") " pod="openshift-marketplace/community-operators-h54gs" Nov 24 17:05:13 crc kubenswrapper[4777]: I1124 17:05:13.091301 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h54gs" Nov 24 17:05:13 crc kubenswrapper[4777]: I1124 17:05:13.111535 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5m4fr"] Nov 24 17:05:13 crc kubenswrapper[4777]: W1124 17:05:13.137942 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6aa2fa6f_bd34_4154_b666_455e7fea71ae.slice/crio-741239a59533c3418c5f77fd8357ab4ff5e8eef9cb81dc0d2bb4b8164be3564b WatchSource:0}: Error finding container 741239a59533c3418c5f77fd8357ab4ff5e8eef9cb81dc0d2bb4b8164be3564b: Status 404 returned error can't find the container with id 741239a59533c3418c5f77fd8357ab4ff5e8eef9cb81dc0d2bb4b8164be3564b Nov 24 17:05:13 crc kubenswrapper[4777]: I1124 17:05:13.275471 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h54gs"] Nov 24 17:05:13 crc kubenswrapper[4777]: I1124 17:05:13.674276 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hwftb" event={"ID":"27733c5c-8536-4f12-8aa0-e5a9d16e8f88","Type":"ContainerStarted","Data":"9a3adf08b07b97b225898f35fb7e3229cf0140f0f34ebce01a960011dcf6d3f2"} Nov 24 17:05:13 crc kubenswrapper[4777]: I1124 17:05:13.676241 4777 generic.go:334] "Generic (PLEG): container finished" podID="6aa2fa6f-bd34-4154-b666-455e7fea71ae" containerID="c91fba508feb987881e0a910d143286938dca7a7b6e049c2e31c50388c0e4fcb" exitCode=0 Nov 24 17:05:13 crc kubenswrapper[4777]: I1124 17:05:13.676315 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5m4fr" event={"ID":"6aa2fa6f-bd34-4154-b666-455e7fea71ae","Type":"ContainerDied","Data":"c91fba508feb987881e0a910d143286938dca7a7b6e049c2e31c50388c0e4fcb"} Nov 24 17:05:13 crc kubenswrapper[4777]: I1124 17:05:13.676345 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5m4fr" event={"ID":"6aa2fa6f-bd34-4154-b666-455e7fea71ae","Type":"ContainerStarted","Data":"741239a59533c3418c5f77fd8357ab4ff5e8eef9cb81dc0d2bb4b8164be3564b"} Nov 24 17:05:13 crc kubenswrapper[4777]: I1124 17:05:13.680273 4777 generic.go:334] "Generic (PLEG): container finished" podID="30623966-678a-4124-bc99-02d4ccfacb85" containerID="7f75f05dcfc8bd786a468502bce4e92fe230d00456f44418402939649ff389f2" exitCode=0 Nov 24 17:05:13 crc kubenswrapper[4777]: I1124 17:05:13.680400 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h54gs" event={"ID":"30623966-678a-4124-bc99-02d4ccfacb85","Type":"ContainerDied","Data":"7f75f05dcfc8bd786a468502bce4e92fe230d00456f44418402939649ff389f2"} Nov 24 17:05:13 crc kubenswrapper[4777]: I1124 17:05:13.680453 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h54gs" event={"ID":"30623966-678a-4124-bc99-02d4ccfacb85","Type":"ContainerStarted","Data":"d50890402303e259848ef39cc18e0578f8aa2c9714e04168c4ccbb7a60f005f7"} Nov 24 17:05:13 crc kubenswrapper[4777]: I1124 17:05:13.695259 4777 generic.go:334] "Generic (PLEG): container finished" podID="1af51c54-3a40-406e-8743-32b0c04d018f" containerID="b9760e7c9db4f3ef0dcdeb811aad48b2d91a20e0cdbf0ac8a7bb7449b9632629" exitCode=0 Nov 24 17:05:13 crc kubenswrapper[4777]: I1124 17:05:13.695482 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7f5l" event={"ID":"1af51c54-3a40-406e-8743-32b0c04d018f","Type":"ContainerDied","Data":"b9760e7c9db4f3ef0dcdeb811aad48b2d91a20e0cdbf0ac8a7bb7449b9632629"} Nov 24 17:05:13 crc kubenswrapper[4777]: I1124 17:05:13.697329 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hwftb" podStartSLOduration=2.020109624 podStartE2EDuration="3.697307566s" podCreationTimestamp="2025-11-24 17:05:10 +0000 UTC" firstStartedPulling="2025-11-24 17:05:11.653920085 +0000 UTC m=+269.813115134" lastFinishedPulling="2025-11-24 17:05:13.331118027 +0000 UTC m=+271.490313076" observedRunningTime="2025-11-24 17:05:13.693243104 +0000 UTC m=+271.852438163" watchObservedRunningTime="2025-11-24 17:05:13.697307566 +0000 UTC m=+271.856502665" Nov 24 17:05:14 crc kubenswrapper[4777]: I1124 17:05:14.702941 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7f5l" event={"ID":"1af51c54-3a40-406e-8743-32b0c04d018f","Type":"ContainerStarted","Data":"d630740879cfc649f79595f0a3681f8cd3184a7b13cb6451027300f3c228b5a6"} Nov 24 17:05:14 crc kubenswrapper[4777]: I1124 17:05:14.704680 4777 generic.go:334] "Generic (PLEG): container finished" podID="6aa2fa6f-bd34-4154-b666-455e7fea71ae" containerID="159ca3aac3f3d12abbea7660c335b2d5794e9cc667058d06f289390e26ad37fc" exitCode=0 Nov 24 17:05:14 crc kubenswrapper[4777]: I1124 17:05:14.704759 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5m4fr" event={"ID":"6aa2fa6f-bd34-4154-b666-455e7fea71ae","Type":"ContainerDied","Data":"159ca3aac3f3d12abbea7660c335b2d5794e9cc667058d06f289390e26ad37fc"} Nov 24 17:05:14 crc kubenswrapper[4777]: I1124 17:05:14.707760 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h54gs" event={"ID":"30623966-678a-4124-bc99-02d4ccfacb85","Type":"ContainerStarted","Data":"63b577342cb7454f47404a183c96201ab1a12715e12d6aa9a6b34e67885594db"} Nov 24 17:05:14 crc kubenswrapper[4777]: I1124 17:05:14.725595 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g7f5l" podStartSLOduration=2.2900977 podStartE2EDuration="4.725578891s" podCreationTimestamp="2025-11-24 17:05:10 +0000 UTC" firstStartedPulling="2025-11-24 17:05:11.650244185 +0000 UTC m=+269.809439234" lastFinishedPulling="2025-11-24 17:05:14.085725376 +0000 UTC m=+272.244920425" observedRunningTime="2025-11-24 17:05:14.72121016 +0000 UTC m=+272.880405209" watchObservedRunningTime="2025-11-24 17:05:14.725578891 +0000 UTC m=+272.884773940" Nov 24 17:05:15 crc kubenswrapper[4777]: I1124 17:05:15.715615 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5m4fr" event={"ID":"6aa2fa6f-bd34-4154-b666-455e7fea71ae","Type":"ContainerStarted","Data":"c3abc850a6cf02a997f6b72e9f8e5e6b50e27bb2be4754678d983ffece33f9c2"} Nov 24 17:05:15 crc kubenswrapper[4777]: I1124 17:05:15.718686 4777 generic.go:334] "Generic (PLEG): container finished" podID="30623966-678a-4124-bc99-02d4ccfacb85" containerID="63b577342cb7454f47404a183c96201ab1a12715e12d6aa9a6b34e67885594db" exitCode=0 Nov 24 17:05:15 crc kubenswrapper[4777]: I1124 17:05:15.718722 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h54gs" event={"ID":"30623966-678a-4124-bc99-02d4ccfacb85","Type":"ContainerDied","Data":"63b577342cb7454f47404a183c96201ab1a12715e12d6aa9a6b34e67885594db"} Nov 24 17:05:15 crc kubenswrapper[4777]: I1124 17:05:15.738703 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5m4fr" podStartSLOduration=2.345643984 podStartE2EDuration="3.738681715s" podCreationTimestamp="2025-11-24 17:05:12 +0000 UTC" firstStartedPulling="2025-11-24 17:05:13.678898266 +0000 UTC m=+271.838093315" lastFinishedPulling="2025-11-24 17:05:15.071935997 +0000 UTC m=+273.231131046" observedRunningTime="2025-11-24 17:05:15.737812029 +0000 UTC m=+273.897007078" watchObservedRunningTime="2025-11-24 17:05:15.738681715 +0000 UTC m=+273.897876764" Nov 24 17:05:17 crc kubenswrapper[4777]: I1124 17:05:17.757168 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h54gs" event={"ID":"30623966-678a-4124-bc99-02d4ccfacb85","Type":"ContainerStarted","Data":"e743d8354c29263f9bdb7f35ae1db3de729df60cd4f83c9ac3b2366e5a499fc1"} Nov 24 17:05:17 crc kubenswrapper[4777]: I1124 17:05:17.781136 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h54gs" podStartSLOduration=3.339477052 podStartE2EDuration="5.781112786s" podCreationTimestamp="2025-11-24 17:05:12 +0000 UTC" firstStartedPulling="2025-11-24 17:05:13.690898584 +0000 UTC m=+271.850093633" lastFinishedPulling="2025-11-24 17:05:16.132534318 +0000 UTC m=+274.291729367" observedRunningTime="2025-11-24 17:05:17.776871909 +0000 UTC m=+275.936066968" watchObservedRunningTime="2025-11-24 17:05:17.781112786 +0000 UTC m=+275.940307875" Nov 24 17:05:20 crc kubenswrapper[4777]: I1124 17:05:20.471067 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hwftb" Nov 24 17:05:20 crc kubenswrapper[4777]: I1124 17:05:20.471587 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hwftb" Nov 24 17:05:20 crc kubenswrapper[4777]: I1124 17:05:20.520817 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hwftb" Nov 24 17:05:20 crc kubenswrapper[4777]: I1124 17:05:20.680297 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g7f5l" Nov 24 17:05:20 crc kubenswrapper[4777]: I1124 17:05:20.680384 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g7f5l" Nov 24 17:05:20 crc kubenswrapper[4777]: I1124 17:05:20.725862 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g7f5l" Nov 24 17:05:20 crc kubenswrapper[4777]: I1124 17:05:20.828900 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g7f5l" Nov 24 17:05:20 crc kubenswrapper[4777]: I1124 17:05:20.831673 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hwftb" Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.143678 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.143934 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.143985 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.146721 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.146949 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.147378 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.155480 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.157119 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.161581 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.169256 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.183263 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.245102 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.248859 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.366484 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 17:05:21 crc kubenswrapper[4777]: W1124 17:05:21.383995 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-ffe3707ae1a78e8426620dfdbc4bd67b3c8fa7af4b4beb70a5ab8f11a763cf47 WatchSource:0}: Error finding container ffe3707ae1a78e8426620dfdbc4bd67b3c8fa7af4b4beb70a5ab8f11a763cf47: Status 404 returned error can't find the container with id ffe3707ae1a78e8426620dfdbc4bd67b3c8fa7af4b4beb70a5ab8f11a763cf47 Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.469696 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:05:21 crc kubenswrapper[4777]: W1124 17:05:21.677382 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-b6c42ff10e905ce099fbb0bd152c196bafdc8c3a361913d3b0a20ea870de1979 WatchSource:0}: Error finding container b6c42ff10e905ce099fbb0bd152c196bafdc8c3a361913d3b0a20ea870de1979: Status 404 returned error can't find the container with id b6c42ff10e905ce099fbb0bd152c196bafdc8c3a361913d3b0a20ea870de1979 Nov 24 17:05:21 crc kubenswrapper[4777]: W1124 17:05:21.747652 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-58d83a71f1d5bfd3d26572db3017a576005adf3e49c0ed6a9be70d73f34598cd WatchSource:0}: Error finding container 58d83a71f1d5bfd3d26572db3017a576005adf3e49c0ed6a9be70d73f34598cd: Status 404 returned error can't find the container with id 58d83a71f1d5bfd3d26572db3017a576005adf3e49c0ed6a9be70d73f34598cd Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.785595 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1de9ae738010c703030ab006c3a83d9bccddfa2ba7bd2f9aea2116b957680f2d"} Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.785646 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ffe3707ae1a78e8426620dfdbc4bd67b3c8fa7af4b4beb70a5ab8f11a763cf47"} Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.787523 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b6c42ff10e905ce099fbb0bd152c196bafdc8c3a361913d3b0a20ea870de1979"} Nov 24 17:05:21 crc kubenswrapper[4777]: I1124 17:05:21.789091 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"58d83a71f1d5bfd3d26572db3017a576005adf3e49c0ed6a9be70d73f34598cd"} Nov 24 17:05:22 crc kubenswrapper[4777]: I1124 17:05:22.795562 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"140eb4a1358b5936fbcae50ceebbcc19df00e55d6300e7391a084b598e50cbd8"} Nov 24 17:05:22 crc kubenswrapper[4777]: I1124 17:05:22.795701 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:05:22 crc kubenswrapper[4777]: I1124 17:05:22.796887 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d72410457048f8dbb9c41b2901fb6de58dba7a9c3292c2fbd0088b1cbac83feb"} Nov 24 17:05:22 crc kubenswrapper[4777]: I1124 17:05:22.929758 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5m4fr" Nov 24 17:05:22 crc kubenswrapper[4777]: I1124 17:05:22.929815 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5m4fr" Nov 24 17:05:22 crc kubenswrapper[4777]: I1124 17:05:22.978224 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5m4fr" Nov 24 17:05:23 crc kubenswrapper[4777]: I1124 17:05:23.091898 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h54gs" Nov 24 17:05:23 crc kubenswrapper[4777]: I1124 17:05:23.092018 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h54gs" Nov 24 17:05:23 crc kubenswrapper[4777]: I1124 17:05:23.139364 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h54gs" Nov 24 17:05:23 crc kubenswrapper[4777]: I1124 17:05:23.837804 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5m4fr" Nov 24 17:05:23 crc kubenswrapper[4777]: I1124 17:05:23.846955 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h54gs" Nov 24 17:05:51 crc kubenswrapper[4777]: I1124 17:05:51.492694 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 17:06:14 crc kubenswrapper[4777]: I1124 17:06:14.270692 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:06:14 crc kubenswrapper[4777]: I1124 17:06:14.271325 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:06:44 crc kubenswrapper[4777]: I1124 17:06:44.270759 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:06:44 crc kubenswrapper[4777]: I1124 17:06:44.271348 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:07:14 crc kubenswrapper[4777]: I1124 17:07:14.270109 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:07:14 crc kubenswrapper[4777]: I1124 17:07:14.270871 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:07:14 crc kubenswrapper[4777]: I1124 17:07:14.270941 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:07:14 crc kubenswrapper[4777]: I1124 17:07:14.272258 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e035704384e89657cd1567ffda73b162c76374285cb18e948ce33a5511ac19b6"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:07:14 crc kubenswrapper[4777]: I1124 17:07:14.272442 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://e035704384e89657cd1567ffda73b162c76374285cb18e948ce33a5511ac19b6" gracePeriod=600 Nov 24 17:07:14 crc kubenswrapper[4777]: I1124 17:07:14.523308 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="e035704384e89657cd1567ffda73b162c76374285cb18e948ce33a5511ac19b6" exitCode=0 Nov 24 17:07:14 crc kubenswrapper[4777]: I1124 17:07:14.523379 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"e035704384e89657cd1567ffda73b162c76374285cb18e948ce33a5511ac19b6"} Nov 24 17:07:14 crc kubenswrapper[4777]: I1124 17:07:14.523640 4777 scope.go:117] "RemoveContainer" containerID="181e0ce2b9ffd88041f95b951fffa94813bb912cf529dc8b64f70818b5500cc3" Nov 24 17:07:15 crc kubenswrapper[4777]: I1124 17:07:15.532624 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"eb08dd4022cfe7bafee35dcfc7cb94f0a8332bb8d48affe8af298f1e887ff112"} Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.157937 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tp2p6"] Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.160632 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.173367 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tp2p6"] Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.355696 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42c34edf-1335-4e70-bae4-04cd20b8bd1d-registry-tls\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.355780 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42c34edf-1335-4e70-bae4-04cd20b8bd1d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.355830 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.355869 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42c34edf-1335-4e70-bae4-04cd20b8bd1d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.355895 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42c34edf-1335-4e70-bae4-04cd20b8bd1d-registry-certificates\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.355919 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqjr5\" (UniqueName: \"kubernetes.io/projected/42c34edf-1335-4e70-bae4-04cd20b8bd1d-kube-api-access-cqjr5\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.356057 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42c34edf-1335-4e70-bae4-04cd20b8bd1d-trusted-ca\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.356115 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42c34edf-1335-4e70-bae4-04cd20b8bd1d-bound-sa-token\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.375740 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.456726 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42c34edf-1335-4e70-bae4-04cd20b8bd1d-registry-certificates\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.456776 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqjr5\" (UniqueName: \"kubernetes.io/projected/42c34edf-1335-4e70-bae4-04cd20b8bd1d-kube-api-access-cqjr5\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.456854 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42c34edf-1335-4e70-bae4-04cd20b8bd1d-trusted-ca\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.456892 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42c34edf-1335-4e70-bae4-04cd20b8bd1d-bound-sa-token\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.456916 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42c34edf-1335-4e70-bae4-04cd20b8bd1d-registry-tls\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.456938 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42c34edf-1335-4e70-bae4-04cd20b8bd1d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.456989 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42c34edf-1335-4e70-bae4-04cd20b8bd1d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.457946 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42c34edf-1335-4e70-bae4-04cd20b8bd1d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.458315 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42c34edf-1335-4e70-bae4-04cd20b8bd1d-registry-certificates\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.458949 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42c34edf-1335-4e70-bae4-04cd20b8bd1d-trusted-ca\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.463549 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42c34edf-1335-4e70-bae4-04cd20b8bd1d-registry-tls\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.464168 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42c34edf-1335-4e70-bae4-04cd20b8bd1d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.475502 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqjr5\" (UniqueName: \"kubernetes.io/projected/42c34edf-1335-4e70-bae4-04cd20b8bd1d-kube-api-access-cqjr5\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.484272 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42c34edf-1335-4e70-bae4-04cd20b8bd1d-bound-sa-token\") pod \"image-registry-66df7c8f76-tp2p6\" (UID: \"42c34edf-1335-4e70-bae4-04cd20b8bd1d\") " pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:15 crc kubenswrapper[4777]: I1124 17:08:15.776332 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:16 crc kubenswrapper[4777]: I1124 17:08:16.075324 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tp2p6"] Nov 24 17:08:16 crc kubenswrapper[4777]: W1124 17:08:16.088663 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42c34edf_1335_4e70_bae4_04cd20b8bd1d.slice/crio-ddf2867f27268dfe79559970ba2d2f3a509b3405ab6e441f4dba76d0ce1a15b4 WatchSource:0}: Error finding container ddf2867f27268dfe79559970ba2d2f3a509b3405ab6e441f4dba76d0ce1a15b4: Status 404 returned error can't find the container with id ddf2867f27268dfe79559970ba2d2f3a509b3405ab6e441f4dba76d0ce1a15b4 Nov 24 17:08:16 crc kubenswrapper[4777]: I1124 17:08:16.940246 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" event={"ID":"42c34edf-1335-4e70-bae4-04cd20b8bd1d","Type":"ContainerStarted","Data":"796d31f99acc6f2a065ea0a3e2475859b5e6273314a52cc58756a79f3e4a4858"} Nov 24 17:08:16 crc kubenswrapper[4777]: I1124 17:08:16.940303 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" event={"ID":"42c34edf-1335-4e70-bae4-04cd20b8bd1d","Type":"ContainerStarted","Data":"ddf2867f27268dfe79559970ba2d2f3a509b3405ab6e441f4dba76d0ce1a15b4"} Nov 24 17:08:16 crc kubenswrapper[4777]: I1124 17:08:16.940440 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:16 crc kubenswrapper[4777]: I1124 17:08:16.967907 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" podStartSLOduration=1.967876384 podStartE2EDuration="1.967876384s" podCreationTimestamp="2025-11-24 17:08:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:08:16.967805852 +0000 UTC m=+455.127000931" watchObservedRunningTime="2025-11-24 17:08:16.967876384 +0000 UTC m=+455.127071473" Nov 24 17:08:35 crc kubenswrapper[4777]: I1124 17:08:35.787379 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-tp2p6" Nov 24 17:08:35 crc kubenswrapper[4777]: I1124 17:08:35.896571 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wsl2k"] Nov 24 17:09:00 crc kubenswrapper[4777]: I1124 17:09:00.948159 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" podUID="1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5" containerName="registry" containerID="cri-o://2ef6bb31c9d4f22edf3828875d522daa9702614edd54f87315bd5488d00e9060" gracePeriod=30 Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.230646 4777 generic.go:334] "Generic (PLEG): container finished" podID="1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5" containerID="2ef6bb31c9d4f22edf3828875d522daa9702614edd54f87315bd5488d00e9060" exitCode=0 Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.230911 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" event={"ID":"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5","Type":"ContainerDied","Data":"2ef6bb31c9d4f22edf3828875d522daa9702614edd54f87315bd5488d00e9060"} Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.324751 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.424351 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-installation-pull-secrets\") pod \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.424413 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-trusted-ca\") pod \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.424493 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-registry-certificates\") pod \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.424692 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.424739 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-ca-trust-extracted\") pod \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.424767 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-registry-tls\") pod \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.424795 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-bound-sa-token\") pod \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.424832 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdjkt\" (UniqueName: \"kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-kube-api-access-wdjkt\") pod \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\" (UID: \"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5\") " Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.425894 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.426078 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.432641 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-kube-api-access-wdjkt" (OuterVolumeSpecName: "kube-api-access-wdjkt") pod "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5"). InnerVolumeSpecName "kube-api-access-wdjkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.432746 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.433206 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.433634 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.440545 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.464930 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5" (UID: "1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.527244 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdjkt\" (UniqueName: \"kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-kube-api-access-wdjkt\") on node \"crc\" DevicePath \"\"" Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.527490 4777 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.527567 4777 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.527671 4777 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.527754 4777 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.527820 4777 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:09:01 crc kubenswrapper[4777]: I1124 17:09:01.527997 4777 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 17:09:02 crc kubenswrapper[4777]: I1124 17:09:02.240025 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" event={"ID":"1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5","Type":"ContainerDied","Data":"0c74eba9d1bc898dc08ec8ab2e8b7219955c87e6511d9bcb08c8de56f946a4d4"} Nov 24 17:09:02 crc kubenswrapper[4777]: I1124 17:09:02.240100 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wsl2k" Nov 24 17:09:02 crc kubenswrapper[4777]: I1124 17:09:02.240111 4777 scope.go:117] "RemoveContainer" containerID="2ef6bb31c9d4f22edf3828875d522daa9702614edd54f87315bd5488d00e9060" Nov 24 17:09:02 crc kubenswrapper[4777]: I1124 17:09:02.291703 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wsl2k"] Nov 24 17:09:02 crc kubenswrapper[4777]: I1124 17:09:02.298432 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wsl2k"] Nov 24 17:09:03 crc kubenswrapper[4777]: I1124 17:09:03.254600 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5" path="/var/lib/kubelet/pods/1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5/volumes" Nov 24 17:09:14 crc kubenswrapper[4777]: I1124 17:09:14.271138 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:09:14 crc kubenswrapper[4777]: I1124 17:09:14.271901 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:09:44 crc kubenswrapper[4777]: I1124 17:09:44.270343 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:09:44 crc kubenswrapper[4777]: I1124 17:09:44.271128 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:10:14 crc kubenswrapper[4777]: I1124 17:10:14.270767 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:10:14 crc kubenswrapper[4777]: I1124 17:10:14.271461 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:10:14 crc kubenswrapper[4777]: I1124 17:10:14.271520 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:10:14 crc kubenswrapper[4777]: I1124 17:10:14.272291 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eb08dd4022cfe7bafee35dcfc7cb94f0a8332bb8d48affe8af298f1e887ff112"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:10:14 crc kubenswrapper[4777]: I1124 17:10:14.272387 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://eb08dd4022cfe7bafee35dcfc7cb94f0a8332bb8d48affe8af298f1e887ff112" gracePeriod=600 Nov 24 17:10:14 crc kubenswrapper[4777]: I1124 17:10:14.725618 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="eb08dd4022cfe7bafee35dcfc7cb94f0a8332bb8d48affe8af298f1e887ff112" exitCode=0 Nov 24 17:10:14 crc kubenswrapper[4777]: I1124 17:10:14.725712 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"eb08dd4022cfe7bafee35dcfc7cb94f0a8332bb8d48affe8af298f1e887ff112"} Nov 24 17:10:14 crc kubenswrapper[4777]: I1124 17:10:14.726176 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"a0f6b6c1a5fc2ed6243c1d079f9f074d4a99e988147412f7b07f358cd4f1c26d"} Nov 24 17:10:14 crc kubenswrapper[4777]: I1124 17:10:14.726211 4777 scope.go:117] "RemoveContainer" containerID="e035704384e89657cd1567ffda73b162c76374285cb18e948ce33a5511ac19b6" Nov 24 17:10:33 crc kubenswrapper[4777]: I1124 17:10:33.752438 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9"] Nov 24 17:10:33 crc kubenswrapper[4777]: E1124 17:10:33.753291 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5" containerName="registry" Nov 24 17:10:33 crc kubenswrapper[4777]: I1124 17:10:33.753311 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5" containerName="registry" Nov 24 17:10:33 crc kubenswrapper[4777]: I1124 17:10:33.753502 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fca2ba6-b5e5-42fd-abe6-a7eb14a0dbc5" containerName="registry" Nov 24 17:10:33 crc kubenswrapper[4777]: I1124 17:10:33.754719 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" Nov 24 17:10:33 crc kubenswrapper[4777]: I1124 17:10:33.756732 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 17:10:33 crc kubenswrapper[4777]: I1124 17:10:33.770703 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9"] Nov 24 17:10:33 crc kubenswrapper[4777]: I1124 17:10:33.821662 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9\" (UID: \"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" Nov 24 17:10:33 crc kubenswrapper[4777]: I1124 17:10:33.821746 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s44fm\" (UniqueName: \"kubernetes.io/projected/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-kube-api-access-s44fm\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9\" (UID: \"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" Nov 24 17:10:33 crc kubenswrapper[4777]: I1124 17:10:33.821799 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9\" (UID: \"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" Nov 24 17:10:33 crc kubenswrapper[4777]: I1124 17:10:33.922930 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9\" (UID: \"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" Nov 24 17:10:33 crc kubenswrapper[4777]: I1124 17:10:33.923089 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9\" (UID: \"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" Nov 24 17:10:33 crc kubenswrapper[4777]: I1124 17:10:33.923184 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s44fm\" (UniqueName: \"kubernetes.io/projected/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-kube-api-access-s44fm\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9\" (UID: \"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" Nov 24 17:10:33 crc kubenswrapper[4777]: I1124 17:10:33.923819 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9\" (UID: \"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" Nov 24 17:10:33 crc kubenswrapper[4777]: I1124 17:10:33.923888 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9\" (UID: \"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" Nov 24 17:10:33 crc kubenswrapper[4777]: I1124 17:10:33.948680 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s44fm\" (UniqueName: \"kubernetes.io/projected/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-kube-api-access-s44fm\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9\" (UID: \"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" Nov 24 17:10:34 crc kubenswrapper[4777]: I1124 17:10:34.074451 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" Nov 24 17:10:34 crc kubenswrapper[4777]: I1124 17:10:34.340799 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9"] Nov 24 17:10:34 crc kubenswrapper[4777]: W1124 17:10:34.351174 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9194629c_e5c7_4ef7_8b08_67ae8d58a1a2.slice/crio-41f8acf0f3372ad6b3eb8ffd742ff36aeb9c89a6171d8fbc0d16712efd7eaf23 WatchSource:0}: Error finding container 41f8acf0f3372ad6b3eb8ffd742ff36aeb9c89a6171d8fbc0d16712efd7eaf23: Status 404 returned error can't find the container with id 41f8acf0f3372ad6b3eb8ffd742ff36aeb9c89a6171d8fbc0d16712efd7eaf23 Nov 24 17:10:34 crc kubenswrapper[4777]: I1124 17:10:34.856736 4777 generic.go:334] "Generic (PLEG): container finished" podID="9194629c-e5c7-4ef7-8b08-67ae8d58a1a2" containerID="51e71edbcd83264685482fd162016a1ebae3c8946c1c4f82d0989a9eefe3ceee" exitCode=0 Nov 24 17:10:34 crc kubenswrapper[4777]: I1124 17:10:34.856809 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" event={"ID":"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2","Type":"ContainerDied","Data":"51e71edbcd83264685482fd162016a1ebae3c8946c1c4f82d0989a9eefe3ceee"} Nov 24 17:10:34 crc kubenswrapper[4777]: I1124 17:10:34.857288 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" event={"ID":"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2","Type":"ContainerStarted","Data":"41f8acf0f3372ad6b3eb8ffd742ff36aeb9c89a6171d8fbc0d16712efd7eaf23"} Nov 24 17:10:34 crc kubenswrapper[4777]: I1124 17:10:34.859236 4777 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:10:36 crc kubenswrapper[4777]: I1124 17:10:36.871169 4777 generic.go:334] "Generic (PLEG): container finished" podID="9194629c-e5c7-4ef7-8b08-67ae8d58a1a2" containerID="5fa1c09c614f907784fddb3610cd0cea12c6c7ced6fed8b2beb029da9bfddc5c" exitCode=0 Nov 24 17:10:36 crc kubenswrapper[4777]: I1124 17:10:36.871261 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" event={"ID":"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2","Type":"ContainerDied","Data":"5fa1c09c614f907784fddb3610cd0cea12c6c7ced6fed8b2beb029da9bfddc5c"} Nov 24 17:10:37 crc kubenswrapper[4777]: I1124 17:10:37.882684 4777 generic.go:334] "Generic (PLEG): container finished" podID="9194629c-e5c7-4ef7-8b08-67ae8d58a1a2" containerID="c11040490c2192f0ed93c2d9a30f6dbd96fb0d827226fe51b55d0c01ed3d21d1" exitCode=0 Nov 24 17:10:37 crc kubenswrapper[4777]: I1124 17:10:37.882787 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" event={"ID":"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2","Type":"ContainerDied","Data":"c11040490c2192f0ed93c2d9a30f6dbd96fb0d827226fe51b55d0c01ed3d21d1"} Nov 24 17:10:39 crc kubenswrapper[4777]: I1124 17:10:39.173067 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" Nov 24 17:10:39 crc kubenswrapper[4777]: I1124 17:10:39.198138 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-bundle\") pod \"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2\" (UID: \"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2\") " Nov 24 17:10:39 crc kubenswrapper[4777]: I1124 17:10:39.198211 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-util\") pod \"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2\" (UID: \"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2\") " Nov 24 17:10:39 crc kubenswrapper[4777]: I1124 17:10:39.198298 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s44fm\" (UniqueName: \"kubernetes.io/projected/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-kube-api-access-s44fm\") pod \"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2\" (UID: \"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2\") " Nov 24 17:10:39 crc kubenswrapper[4777]: I1124 17:10:39.210805 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-bundle" (OuterVolumeSpecName: "bundle") pod "9194629c-e5c7-4ef7-8b08-67ae8d58a1a2" (UID: "9194629c-e5c7-4ef7-8b08-67ae8d58a1a2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:10:39 crc kubenswrapper[4777]: I1124 17:10:39.212426 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-kube-api-access-s44fm" (OuterVolumeSpecName: "kube-api-access-s44fm") pod "9194629c-e5c7-4ef7-8b08-67ae8d58a1a2" (UID: "9194629c-e5c7-4ef7-8b08-67ae8d58a1a2"). InnerVolumeSpecName "kube-api-access-s44fm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:10:39 crc kubenswrapper[4777]: I1124 17:10:39.233969 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-util" (OuterVolumeSpecName: "util") pod "9194629c-e5c7-4ef7-8b08-67ae8d58a1a2" (UID: "9194629c-e5c7-4ef7-8b08-67ae8d58a1a2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:10:39 crc kubenswrapper[4777]: I1124 17:10:39.299410 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s44fm\" (UniqueName: \"kubernetes.io/projected/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-kube-api-access-s44fm\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:39 crc kubenswrapper[4777]: I1124 17:10:39.299456 4777 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:39 crc kubenswrapper[4777]: I1124 17:10:39.299472 4777 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9194629c-e5c7-4ef7-8b08-67ae8d58a1a2-util\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:39 crc kubenswrapper[4777]: I1124 17:10:39.902231 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" event={"ID":"9194629c-e5c7-4ef7-8b08-67ae8d58a1a2","Type":"ContainerDied","Data":"41f8acf0f3372ad6b3eb8ffd742ff36aeb9c89a6171d8fbc0d16712efd7eaf23"} Nov 24 17:10:39 crc kubenswrapper[4777]: I1124 17:10:39.902302 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41f8acf0f3372ad6b3eb8ffd742ff36aeb9c89a6171d8fbc0d16712efd7eaf23" Nov 24 17:10:39 crc kubenswrapper[4777]: I1124 17:10:39.902338 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.153961 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-djshf"] Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.154622 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovn-controller" containerID="cri-o://19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366" gracePeriod=30 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.154944 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="sbdb" containerID="cri-o://7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936" gracePeriod=30 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.154995 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="nbdb" containerID="cri-o://a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a" gracePeriod=30 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.155024 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="northd" containerID="cri-o://60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860" gracePeriod=30 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.155055 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab" gracePeriod=30 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.155080 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="kube-rbac-proxy-node" containerID="cri-o://6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807" gracePeriod=30 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.155106 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovn-acl-logging" containerID="cri-o://c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b" gracePeriod=30 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.221225 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" containerID="cri-o://8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605" gracePeriod=30 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.505756 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/3.log" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.507754 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovn-acl-logging/0.log" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.508449 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovn-controller/0.log" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.508825 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.573773 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gxcwh"] Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.574165 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="nbdb" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.574235 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="nbdb" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.574283 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9194629c-e5c7-4ef7-8b08-67ae8d58a1a2" containerName="extract" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.574330 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9194629c-e5c7-4ef7-8b08-67ae8d58a1a2" containerName="extract" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.574376 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.574440 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.574485 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="kubecfg-setup" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.574533 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="kubecfg-setup" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.574579 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.574621 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.574670 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.574713 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.574755 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.574798 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.574851 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9194629c-e5c7-4ef7-8b08-67ae8d58a1a2" containerName="pull" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.574898 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9194629c-e5c7-4ef7-8b08-67ae8d58a1a2" containerName="pull" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.574943 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.575002 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.575052 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovn-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.575094 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovn-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.575149 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9194629c-e5c7-4ef7-8b08-67ae8d58a1a2" containerName="util" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.575192 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9194629c-e5c7-4ef7-8b08-67ae8d58a1a2" containerName="util" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.575238 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="sbdb" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.575280 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="sbdb" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.575327 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="northd" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.575368 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="northd" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.575418 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="kube-rbac-proxy-node" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.575461 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="kube-rbac-proxy-node" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.575505 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovn-acl-logging" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.575551 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovn-acl-logging" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.575673 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.575723 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9194629c-e5c7-4ef7-8b08-67ae8d58a1a2" containerName="extract" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.575767 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovn-acl-logging" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.575821 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovn-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.575889 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="nbdb" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.575936 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.576000 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.576046 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.576089 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="northd" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.576153 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="kube-rbac-proxy-node" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.576200 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.576244 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="sbdb" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.576448 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.576500 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.576628 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerName="ovnkube-controller" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.577243 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-cni-netd\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.577322 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.577350 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovnkube-script-lib\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.577500 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-ovn\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.577646 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-log-socket\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.577768 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-openvswitch\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.577864 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-run-ovn-kubernetes\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.577954 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-cni-bin\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.578099 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovn-node-metrics-cert\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.578875 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-var-lib-cni-networks-ovn-kubernetes\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.578963 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-etc-openvswitch\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.579058 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-slash\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.577598 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.577727 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-log-socket" (OuterVolumeSpecName: "log-socket") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.577835 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.577840 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.577928 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.578065 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.578956 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.579056 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.579096 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-slash" (OuterVolumeSpecName: "host-slash") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.579271 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.579117 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-run-netns\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.579451 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-systemd\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.579518 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-node-log\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.579591 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-kubelet\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.579658 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovnkube-config\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.579725 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-systemd-units\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.579788 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kk9ks\" (UniqueName: \"kubernetes.io/projected/9084c33d-bd72-470c-9f45-3b0b0f29aa19-kube-api-access-kk9ks\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.579845 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-var-lib-openvswitch\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.580004 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-env-overrides\") pod \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\" (UID: \"9084c33d-bd72-470c-9f45-3b0b0f29aa19\") " Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.579545 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-node-log" (OuterVolumeSpecName: "node-log") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.579690 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.580092 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.580118 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.580221 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.580439 4777 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.580607 4777 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.580660 4777 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-log-socket\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.580711 4777 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.580764 4777 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.580812 4777 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.580857 4777 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.580908 4777 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.580954 4777 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-slash\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.581018 4777 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.581077 4777 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-node-log\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.581135 4777 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.581188 4777 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.581239 4777 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.581285 4777 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.581330 4777 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.580456 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.582681 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.588499 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9084c33d-bd72-470c-9f45-3b0b0f29aa19-kube-api-access-kk9ks" (OuterVolumeSpecName: "kube-api-access-kk9ks") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "kube-api-access-kk9ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.589762 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.591961 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "9084c33d-bd72-470c-9f45-3b0b0f29aa19" (UID: "9084c33d-bd72-470c-9f45-3b0b0f29aa19"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.682388 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-run-systemd\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.682669 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-ovnkube-config\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.682738 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-var-lib-openvswitch\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.682801 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-run-ovn\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.682856 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-kubelet\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.682947 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-log-socket\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.683031 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-env-overrides\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.683100 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-systemd-units\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.683160 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-node-log\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.683224 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-run-openvswitch\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.683292 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-etc-openvswitch\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.683361 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-run-ovn-kubernetes\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.683418 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-cni-netd\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.683532 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-run-netns\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.683591 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-slash\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.683654 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.683714 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4twp\" (UniqueName: \"kubernetes.io/projected/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-kube-api-access-v4twp\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.683779 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-ovn-node-metrics-cert\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.683835 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-ovnkube-script-lib\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.683895 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-cni-bin\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.683988 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kk9ks\" (UniqueName: \"kubernetes.io/projected/9084c33d-bd72-470c-9f45-3b0b0f29aa19-kube-api-access-kk9ks\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.684043 4777 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9084c33d-bd72-470c-9f45-3b0b0f29aa19-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.684092 4777 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9084c33d-bd72-470c-9f45-3b0b0f29aa19-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.684191 4777 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9084c33d-bd72-470c-9f45-3b0b0f29aa19-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.785161 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-run-ovn-kubernetes\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.785407 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-cni-netd\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.785470 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-run-netns\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.785530 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-slash\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.785597 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.785656 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4twp\" (UniqueName: \"kubernetes.io/projected/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-kube-api-access-v4twp\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.785715 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-ovnkube-script-lib\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.785777 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-ovn-node-metrics-cert\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.785836 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-cni-bin\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.785900 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-run-systemd\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.785990 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-ovnkube-config\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.786059 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-var-lib-openvswitch\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.786130 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-run-ovn\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.786188 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-kubelet\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.786244 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-log-socket\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.786312 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-env-overrides\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.786373 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-systemd-units\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.786432 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-node-log\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.786488 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-run-openvswitch\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.786548 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-etc-openvswitch\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.786666 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-etc-openvswitch\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.785236 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-run-ovn-kubernetes\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.786798 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-cni-netd\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.786864 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-run-netns\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.786929 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-slash\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.787020 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.787723 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-run-ovn\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.787920 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-ovnkube-script-lib\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.788020 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-kubelet\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.788091 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-log-socket\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.788455 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-env-overrides\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.788544 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-systemd-units\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.788611 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-node-log\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.788679 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-run-openvswitch\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.788746 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-run-systemd\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.788816 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-host-cni-bin\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.789251 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-ovnkube-config\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.789339 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-var-lib-openvswitch\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.791050 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-ovn-node-metrics-cert\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.809220 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4twp\" (UniqueName: \"kubernetes.io/projected/0f25f5a5-fdf6-4c44-9849-ece6ef97b698-kube-api-access-v4twp\") pod \"ovnkube-node-gxcwh\" (UID: \"0f25f5a5-fdf6-4c44-9849-ece6ef97b698\") " pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.899370 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.931330 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" event={"ID":"0f25f5a5-fdf6-4c44-9849-ece6ef97b698","Type":"ContainerStarted","Data":"3c418cfd5c20408262be37f685d59f0307cb7a370652369f7602853171c5db95"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.934062 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovnkube-controller/3.log" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.936996 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovn-acl-logging/0.log" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.937500 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-djshf_9084c33d-bd72-470c-9f45-3b0b0f29aa19/ovn-controller/0.log" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.937901 4777 generic.go:334] "Generic (PLEG): container finished" podID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerID="8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605" exitCode=0 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.937998 4777 generic.go:334] "Generic (PLEG): container finished" podID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerID="7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936" exitCode=0 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938054 4777 generic.go:334] "Generic (PLEG): container finished" podID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerID="a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a" exitCode=0 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938100 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938111 4777 generic.go:334] "Generic (PLEG): container finished" podID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerID="60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860" exitCode=0 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938294 4777 generic.go:334] "Generic (PLEG): container finished" podID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerID="0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab" exitCode=0 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938332 4777 generic.go:334] "Generic (PLEG): container finished" podID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerID="6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807" exitCode=0 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938351 4777 generic.go:334] "Generic (PLEG): container finished" podID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerID="c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b" exitCode=143 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938367 4777 generic.go:334] "Generic (PLEG): container finished" podID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" containerID="19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366" exitCode=143 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938008 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerDied","Data":"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938529 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerDied","Data":"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938602 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerDied","Data":"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938673 4777 scope.go:117] "RemoveContainer" containerID="8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938676 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerDied","Data":"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938830 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerDied","Data":"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938866 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerDied","Data":"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938891 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938904 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938910 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938916 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938922 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938928 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938933 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938939 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938945 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938954 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerDied","Data":"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.938998 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939006 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939012 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939017 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939022 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939027 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939031 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939036 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939041 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939046 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939053 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerDied","Data":"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939062 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939068 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939076 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939081 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939086 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939091 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939096 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939101 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939106 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939112 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939120 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-djshf" event={"ID":"9084c33d-bd72-470c-9f45-3b0b0f29aa19","Type":"ContainerDied","Data":"93a349ce411ab44f6eb3c91d06db3e86b9dc6d7413f6af0b8575c8c9c0253ffd"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939127 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939133 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939138 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939146 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939152 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939157 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939162 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939168 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939174 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.939179 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.942082 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mfvtr_a5500f09-0e35-442c-87ae-8f280cd2edd1/kube-multus/2.log" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.945540 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mfvtr_a5500f09-0e35-442c-87ae-8f280cd2edd1/kube-multus/1.log" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.945627 4777 generic.go:334] "Generic (PLEG): container finished" podID="a5500f09-0e35-442c-87ae-8f280cd2edd1" containerID="f481bf73ae67258e05eff97b7b2653b611d0cd13cac8da9e0655f728f5bc954f" exitCode=2 Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.945672 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mfvtr" event={"ID":"a5500f09-0e35-442c-87ae-8f280cd2edd1","Type":"ContainerDied","Data":"f481bf73ae67258e05eff97b7b2653b611d0cd13cac8da9e0655f728f5bc954f"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.945773 4777 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"899132c850ac1d7af915f5cb557c347ff8caec6f210b43743788369f5852c7e8"} Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.946464 4777 scope.go:117] "RemoveContainer" containerID="f481bf73ae67258e05eff97b7b2653b611d0cd13cac8da9e0655f728f5bc954f" Nov 24 17:10:45 crc kubenswrapper[4777]: E1124 17:10:45.946751 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-mfvtr_openshift-multus(a5500f09-0e35-442c-87ae-8f280cd2edd1)\"" pod="openshift-multus/multus-mfvtr" podUID="a5500f09-0e35-442c-87ae-8f280cd2edd1" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.969946 4777 scope.go:117] "RemoveContainer" containerID="1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e" Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.991062 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-djshf"] Nov 24 17:10:45 crc kubenswrapper[4777]: I1124 17:10:45.992792 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-djshf"] Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.015988 4777 scope.go:117] "RemoveContainer" containerID="7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.044965 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx"] Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.045578 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.046629 4777 scope.go:117] "RemoveContainer" containerID="a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.047336 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-9z5bk" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.047717 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.048651 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.063208 4777 scope.go:117] "RemoveContainer" containerID="60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.081427 4777 scope.go:117] "RemoveContainer" containerID="0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.090216 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98gkp\" (UniqueName: \"kubernetes.io/projected/b08a2563-7e34-40f7-b9b0-052824422cff-kube-api-access-98gkp\") pod \"obo-prometheus-operator-668cf9dfbb-wsbfx\" (UID: \"b08a2563-7e34-40f7-b9b0-052824422cff\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.095073 4777 scope.go:117] "RemoveContainer" containerID="6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.149091 4777 scope.go:117] "RemoveContainer" containerID="c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.163619 4777 scope.go:117] "RemoveContainer" containerID="19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.177743 4777 scope.go:117] "RemoveContainer" containerID="94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.191559 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98gkp\" (UniqueName: \"kubernetes.io/projected/b08a2563-7e34-40f7-b9b0-052824422cff-kube-api-access-98gkp\") pod \"obo-prometheus-operator-668cf9dfbb-wsbfx\" (UID: \"b08a2563-7e34-40f7-b9b0-052824422cff\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.195857 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq"] Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.196853 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.197183 4777 scope.go:117] "RemoveContainer" containerID="8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.198372 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605\": container with ID starting with 8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605 not found: ID does not exist" containerID="8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.198421 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605"} err="failed to get container status \"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605\": rpc error: code = NotFound desc = could not find container \"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605\": container with ID starting with 8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.198449 4777 scope.go:117] "RemoveContainer" containerID="1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.199210 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\": container with ID starting with 1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e not found: ID does not exist" containerID="1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.199312 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e"} err="failed to get container status \"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\": rpc error: code = NotFound desc = could not find container \"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\": container with ID starting with 1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.199392 4777 scope.go:117] "RemoveContainer" containerID="7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.203595 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\": container with ID starting with 7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936 not found: ID does not exist" containerID="7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.203646 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936"} err="failed to get container status \"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\": rpc error: code = NotFound desc = could not find container \"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\": container with ID starting with 7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.203677 4777 scope.go:117] "RemoveContainer" containerID="a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.204088 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\": container with ID starting with a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a not found: ID does not exist" containerID="a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.204131 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a"} err="failed to get container status \"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\": rpc error: code = NotFound desc = could not find container \"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\": container with ID starting with a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.204161 4777 scope.go:117] "RemoveContainer" containerID="60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.204485 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\": container with ID starting with 60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860 not found: ID does not exist" containerID="60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.204551 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860"} err="failed to get container status \"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\": rpc error: code = NotFound desc = could not find container \"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\": container with ID starting with 60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.204608 4777 scope.go:117] "RemoveContainer" containerID="0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.204616 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-pqm4k" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.204937 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\": container with ID starting with 0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab not found: ID does not exist" containerID="0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.204982 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab"} err="failed to get container status \"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\": rpc error: code = NotFound desc = could not find container \"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\": container with ID starting with 0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.205001 4777 scope.go:117] "RemoveContainer" containerID="6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.205151 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.205482 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\": container with ID starting with 6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807 not found: ID does not exist" containerID="6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.205504 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807"} err="failed to get container status \"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\": rpc error: code = NotFound desc = could not find container \"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\": container with ID starting with 6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.205521 4777 scope.go:117] "RemoveContainer" containerID="c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.206235 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\": container with ID starting with c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b not found: ID does not exist" containerID="c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.206307 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b"} err="failed to get container status \"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\": rpc error: code = NotFound desc = could not find container \"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\": container with ID starting with c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.206378 4777 scope.go:117] "RemoveContainer" containerID="19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.206675 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\": container with ID starting with 19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366 not found: ID does not exist" containerID="19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.206708 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366"} err="failed to get container status \"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\": rpc error: code = NotFound desc = could not find container \"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\": container with ID starting with 19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.206728 4777 scope.go:117] "RemoveContainer" containerID="94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.207022 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\": container with ID starting with 94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc not found: ID does not exist" containerID="94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.207057 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc"} err="failed to get container status \"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\": rpc error: code = NotFound desc = could not find container \"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\": container with ID starting with 94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.207083 4777 scope.go:117] "RemoveContainer" containerID="8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.207274 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r"] Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.208094 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.210433 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605"} err="failed to get container status \"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605\": rpc error: code = NotFound desc = could not find container \"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605\": container with ID starting with 8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.210512 4777 scope.go:117] "RemoveContainer" containerID="1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.210861 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e"} err="failed to get container status \"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\": rpc error: code = NotFound desc = could not find container \"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\": container with ID starting with 1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.210905 4777 scope.go:117] "RemoveContainer" containerID="7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.211281 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936"} err="failed to get container status \"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\": rpc error: code = NotFound desc = could not find container \"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\": container with ID starting with 7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.211310 4777 scope.go:117] "RemoveContainer" containerID="a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.211659 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a"} err="failed to get container status \"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\": rpc error: code = NotFound desc = could not find container \"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\": container with ID starting with a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.211679 4777 scope.go:117] "RemoveContainer" containerID="60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.212009 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860"} err="failed to get container status \"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\": rpc error: code = NotFound desc = could not find container \"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\": container with ID starting with 60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.212028 4777 scope.go:117] "RemoveContainer" containerID="0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.212313 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab"} err="failed to get container status \"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\": rpc error: code = NotFound desc = could not find container \"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\": container with ID starting with 0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.212347 4777 scope.go:117] "RemoveContainer" containerID="6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.212598 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807"} err="failed to get container status \"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\": rpc error: code = NotFound desc = could not find container \"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\": container with ID starting with 6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.212617 4777 scope.go:117] "RemoveContainer" containerID="c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.212825 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b"} err="failed to get container status \"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\": rpc error: code = NotFound desc = could not find container \"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\": container with ID starting with c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.212839 4777 scope.go:117] "RemoveContainer" containerID="19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.213053 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366"} err="failed to get container status \"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\": rpc error: code = NotFound desc = could not find container \"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\": container with ID starting with 19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.213068 4777 scope.go:117] "RemoveContainer" containerID="94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.213892 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc"} err="failed to get container status \"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\": rpc error: code = NotFound desc = could not find container \"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\": container with ID starting with 94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.213915 4777 scope.go:117] "RemoveContainer" containerID="8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.214190 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605"} err="failed to get container status \"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605\": rpc error: code = NotFound desc = could not find container \"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605\": container with ID starting with 8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.214204 4777 scope.go:117] "RemoveContainer" containerID="1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.214390 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e"} err="failed to get container status \"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\": rpc error: code = NotFound desc = could not find container \"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\": container with ID starting with 1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.214402 4777 scope.go:117] "RemoveContainer" containerID="7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.214644 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936"} err="failed to get container status \"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\": rpc error: code = NotFound desc = could not find container \"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\": container with ID starting with 7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.214660 4777 scope.go:117] "RemoveContainer" containerID="a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.214913 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a"} err="failed to get container status \"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\": rpc error: code = NotFound desc = could not find container \"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\": container with ID starting with a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.214926 4777 scope.go:117] "RemoveContainer" containerID="60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.215270 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860"} err="failed to get container status \"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\": rpc error: code = NotFound desc = could not find container \"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\": container with ID starting with 60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.215307 4777 scope.go:117] "RemoveContainer" containerID="0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.215731 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab"} err="failed to get container status \"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\": rpc error: code = NotFound desc = could not find container \"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\": container with ID starting with 0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.215755 4777 scope.go:117] "RemoveContainer" containerID="6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.216073 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807"} err="failed to get container status \"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\": rpc error: code = NotFound desc = could not find container \"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\": container with ID starting with 6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.216090 4777 scope.go:117] "RemoveContainer" containerID="c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.216834 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b"} err="failed to get container status \"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\": rpc error: code = NotFound desc = could not find container \"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\": container with ID starting with c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.216855 4777 scope.go:117] "RemoveContainer" containerID="19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.217122 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366"} err="failed to get container status \"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\": rpc error: code = NotFound desc = could not find container \"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\": container with ID starting with 19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.217148 4777 scope.go:117] "RemoveContainer" containerID="94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.217387 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc"} err="failed to get container status \"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\": rpc error: code = NotFound desc = could not find container \"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\": container with ID starting with 94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.217408 4777 scope.go:117] "RemoveContainer" containerID="8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.217687 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605"} err="failed to get container status \"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605\": rpc error: code = NotFound desc = could not find container \"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605\": container with ID starting with 8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.217713 4777 scope.go:117] "RemoveContainer" containerID="1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.217911 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e"} err="failed to get container status \"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\": rpc error: code = NotFound desc = could not find container \"1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e\": container with ID starting with 1bb13c1744c6de698ac0fd3b2db8a8cf44349428fd8c71f60fd08568188ab04e not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.217947 4777 scope.go:117] "RemoveContainer" containerID="7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.218164 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936"} err="failed to get container status \"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\": rpc error: code = NotFound desc = could not find container \"7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936\": container with ID starting with 7ac6f8ae1daf61c0329a9a1a2003ac8ddda514486c788fc6d3aab91af8742936 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.218183 4777 scope.go:117] "RemoveContainer" containerID="a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.218436 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a"} err="failed to get container status \"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\": rpc error: code = NotFound desc = could not find container \"a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a\": container with ID starting with a6ed4ca7c5896c0b8938d66c7a4613e00e8261804d2d4813a6748724341e3d8a not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.218453 4777 scope.go:117] "RemoveContainer" containerID="60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.218651 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860"} err="failed to get container status \"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\": rpc error: code = NotFound desc = could not find container \"60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860\": container with ID starting with 60361d867e7d8bb9cd5501d76f78747f8921bb1644c96bc048e12b36301ff860 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.218665 4777 scope.go:117] "RemoveContainer" containerID="0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.220538 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab"} err="failed to get container status \"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\": rpc error: code = NotFound desc = could not find container \"0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab\": container with ID starting with 0a0094823e34088f31111c0dbca54cfdbdbb8aed7873eef4cc33c60680c7b4ab not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.220666 4777 scope.go:117] "RemoveContainer" containerID="6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.221387 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98gkp\" (UniqueName: \"kubernetes.io/projected/b08a2563-7e34-40f7-b9b0-052824422cff-kube-api-access-98gkp\") pod \"obo-prometheus-operator-668cf9dfbb-wsbfx\" (UID: \"b08a2563-7e34-40f7-b9b0-052824422cff\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.221694 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807"} err="failed to get container status \"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\": rpc error: code = NotFound desc = could not find container \"6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807\": container with ID starting with 6e5959cf531f4fc1c771dd73e1f0dc86f28ead3b132e4922214a001ce73a1807 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.221738 4777 scope.go:117] "RemoveContainer" containerID="c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.223740 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b"} err="failed to get container status \"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\": rpc error: code = NotFound desc = could not find container \"c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b\": container with ID starting with c6c3febabb52c1d91f1629ccc85e22d1a71247716295478bbb29f87604dfb04b not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.223768 4777 scope.go:117] "RemoveContainer" containerID="19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.224105 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366"} err="failed to get container status \"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\": rpc error: code = NotFound desc = could not find container \"19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366\": container with ID starting with 19aacde00b5557ddb6e1d91224ce11a48ba82395f8f803d6d7d9b68f348c4366 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.224139 4777 scope.go:117] "RemoveContainer" containerID="94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.224394 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc"} err="failed to get container status \"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\": rpc error: code = NotFound desc = could not find container \"94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc\": container with ID starting with 94c7b94f58b33c6d3948ca2183948c62a63c3369e9cc93afd70563f000b683cc not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.224412 4777 scope.go:117] "RemoveContainer" containerID="8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.225529 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605"} err="failed to get container status \"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605\": rpc error: code = NotFound desc = could not find container \"8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605\": container with ID starting with 8a6949b6cc8b471427d26771ed821d51b90a95432df4cb595301e3637dab7605 not found: ID does not exist" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.294297 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dd6addca-3fb6-4204-ae4e-2ed8d1326699-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r\" (UID: \"dd6addca-3fb6-4204-ae4e-2ed8d1326699\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.294390 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/67418bcd-f0ba-4526-afc1-8420cb642ccf-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq\" (UID: \"67418bcd-f0ba-4526-afc1-8420cb642ccf\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.294464 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dd6addca-3fb6-4204-ae4e-2ed8d1326699-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r\" (UID: \"dd6addca-3fb6-4204-ae4e-2ed8d1326699\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.294489 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/67418bcd-f0ba-4526-afc1-8420cb642ccf-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq\" (UID: \"67418bcd-f0ba-4526-afc1-8420cb642ccf\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.359909 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.381851 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators_b08a2563-7e34-40f7-b9b0-052824422cff_0(7845be96b81b488b0c3f2bd0d5de495777dd83f72af4c4bca8016e2ad89cfd91): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.381917 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators_b08a2563-7e34-40f7-b9b0-052824422cff_0(7845be96b81b488b0c3f2bd0d5de495777dd83f72af4c4bca8016e2ad89cfd91): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.381944 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators_b08a2563-7e34-40f7-b9b0-052824422cff_0(7845be96b81b488b0c3f2bd0d5de495777dd83f72af4c4bca8016e2ad89cfd91): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.382020 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators(b08a2563-7e34-40f7-b9b0-052824422cff)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators(b08a2563-7e34-40f7-b9b0-052824422cff)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators_b08a2563-7e34-40f7-b9b0-052824422cff_0(7845be96b81b488b0c3f2bd0d5de495777dd83f72af4c4bca8016e2ad89cfd91): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" podUID="b08a2563-7e34-40f7-b9b0-052824422cff" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.394243 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-hdxt8"] Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.394804 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.395289 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dd6addca-3fb6-4204-ae4e-2ed8d1326699-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r\" (UID: \"dd6addca-3fb6-4204-ae4e-2ed8d1326699\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.395355 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/67418bcd-f0ba-4526-afc1-8420cb642ccf-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq\" (UID: \"67418bcd-f0ba-4526-afc1-8420cb642ccf\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.395419 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dd6addca-3fb6-4204-ae4e-2ed8d1326699-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r\" (UID: \"dd6addca-3fb6-4204-ae4e-2ed8d1326699\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.395452 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/67418bcd-f0ba-4526-afc1-8420cb642ccf-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq\" (UID: \"67418bcd-f0ba-4526-afc1-8420cb642ccf\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.399359 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.399366 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-crmpg" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.399822 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dd6addca-3fb6-4204-ae4e-2ed8d1326699-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r\" (UID: \"dd6addca-3fb6-4204-ae4e-2ed8d1326699\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.401081 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/67418bcd-f0ba-4526-afc1-8420cb642ccf-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq\" (UID: \"67418bcd-f0ba-4526-afc1-8420cb642ccf\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.402513 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dd6addca-3fb6-4204-ae4e-2ed8d1326699-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r\" (UID: \"dd6addca-3fb6-4204-ae4e-2ed8d1326699\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.403368 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/67418bcd-f0ba-4526-afc1-8420cb642ccf-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq\" (UID: \"67418bcd-f0ba-4526-afc1-8420cb642ccf\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.479961 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-22cbh"] Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.480709 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.491264 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-6hpbr" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.496575 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qggr\" (UniqueName: \"kubernetes.io/projected/986d062f-3210-471c-a2db-07e080b5d449-kube-api-access-2qggr\") pod \"observability-operator-d8bb48f5d-hdxt8\" (UID: \"986d062f-3210-471c-a2db-07e080b5d449\") " pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.496698 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c0fe98e7-9ad1-426e-a17a-ad471a7f73af-openshift-service-ca\") pod \"perses-operator-5446b9c989-22cbh\" (UID: \"c0fe98e7-9ad1-426e-a17a-ad471a7f73af\") " pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.496770 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/986d062f-3210-471c-a2db-07e080b5d449-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-hdxt8\" (UID: \"986d062f-3210-471c-a2db-07e080b5d449\") " pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.496850 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6vhp\" (UniqueName: \"kubernetes.io/projected/c0fe98e7-9ad1-426e-a17a-ad471a7f73af-kube-api-access-v6vhp\") pod \"perses-operator-5446b9c989-22cbh\" (UID: \"c0fe98e7-9ad1-426e-a17a-ad471a7f73af\") " pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.514176 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.525427 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.562725 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators_67418bcd-f0ba-4526-afc1-8420cb642ccf_0(7f9cf38b2cc6fb1b0bf0599734609be8820f679126c4082155e06d20ce491632): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.562875 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators_67418bcd-f0ba-4526-afc1-8420cb642ccf_0(7f9cf38b2cc6fb1b0bf0599734609be8820f679126c4082155e06d20ce491632): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.562987 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators_67418bcd-f0ba-4526-afc1-8420cb642ccf_0(7f9cf38b2cc6fb1b0bf0599734609be8820f679126c4082155e06d20ce491632): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.563117 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators(67418bcd-f0ba-4526-afc1-8420cb642ccf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators(67418bcd-f0ba-4526-afc1-8420cb642ccf)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators_67418bcd-f0ba-4526-afc1-8420cb642ccf_0(7f9cf38b2cc6fb1b0bf0599734609be8820f679126c4082155e06d20ce491632): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" podUID="67418bcd-f0ba-4526-afc1-8420cb642ccf" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.576658 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(dd01497cb21c4647596b54cf0ee5e3cdaa415a423993be9311dd160a446a5715): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.576787 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(dd01497cb21c4647596b54cf0ee5e3cdaa415a423993be9311dd160a446a5715): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.576855 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(dd01497cb21c4647596b54cf0ee5e3cdaa415a423993be9311dd160a446a5715): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.576940 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators(dd6addca-3fb6-4204-ae4e-2ed8d1326699)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators(dd6addca-3fb6-4204-ae4e-2ed8d1326699)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(dd01497cb21c4647596b54cf0ee5e3cdaa415a423993be9311dd160a446a5715): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" podUID="dd6addca-3fb6-4204-ae4e-2ed8d1326699" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.598204 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6vhp\" (UniqueName: \"kubernetes.io/projected/c0fe98e7-9ad1-426e-a17a-ad471a7f73af-kube-api-access-v6vhp\") pod \"perses-operator-5446b9c989-22cbh\" (UID: \"c0fe98e7-9ad1-426e-a17a-ad471a7f73af\") " pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.598666 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qggr\" (UniqueName: \"kubernetes.io/projected/986d062f-3210-471c-a2db-07e080b5d449-kube-api-access-2qggr\") pod \"observability-operator-d8bb48f5d-hdxt8\" (UID: \"986d062f-3210-471c-a2db-07e080b5d449\") " pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.598753 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c0fe98e7-9ad1-426e-a17a-ad471a7f73af-openshift-service-ca\") pod \"perses-operator-5446b9c989-22cbh\" (UID: \"c0fe98e7-9ad1-426e-a17a-ad471a7f73af\") " pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.599603 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/986d062f-3210-471c-a2db-07e080b5d449-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-hdxt8\" (UID: \"986d062f-3210-471c-a2db-07e080b5d449\") " pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.599565 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c0fe98e7-9ad1-426e-a17a-ad471a7f73af-openshift-service-ca\") pod \"perses-operator-5446b9c989-22cbh\" (UID: \"c0fe98e7-9ad1-426e-a17a-ad471a7f73af\") " pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.605628 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/986d062f-3210-471c-a2db-07e080b5d449-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-hdxt8\" (UID: \"986d062f-3210-471c-a2db-07e080b5d449\") " pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.618463 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6vhp\" (UniqueName: \"kubernetes.io/projected/c0fe98e7-9ad1-426e-a17a-ad471a7f73af-kube-api-access-v6vhp\") pod \"perses-operator-5446b9c989-22cbh\" (UID: \"c0fe98e7-9ad1-426e-a17a-ad471a7f73af\") " pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.621856 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qggr\" (UniqueName: \"kubernetes.io/projected/986d062f-3210-471c-a2db-07e080b5d449-kube-api-access-2qggr\") pod \"observability-operator-d8bb48f5d-hdxt8\" (UID: \"986d062f-3210-471c-a2db-07e080b5d449\") " pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.742424 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.758937 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-hdxt8_openshift-operators_986d062f-3210-471c-a2db-07e080b5d449_0(88726c89449b098c8caefe073350c20eba44e359d6194346df8f915acec850b2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.759124 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-hdxt8_openshift-operators_986d062f-3210-471c-a2db-07e080b5d449_0(88726c89449b098c8caefe073350c20eba44e359d6194346df8f915acec850b2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.759200 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-hdxt8_openshift-operators_986d062f-3210-471c-a2db-07e080b5d449_0(88726c89449b098c8caefe073350c20eba44e359d6194346df8f915acec850b2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.759289 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-hdxt8_openshift-operators(986d062f-3210-471c-a2db-07e080b5d449)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-hdxt8_openshift-operators(986d062f-3210-471c-a2db-07e080b5d449)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-hdxt8_openshift-operators_986d062f-3210-471c-a2db-07e080b5d449_0(88726c89449b098c8caefe073350c20eba44e359d6194346df8f915acec850b2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" podUID="986d062f-3210-471c-a2db-07e080b5d449" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.792768 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.809450 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(93487ffefc09be98021c5453fb6a2dc904b9b8a5e9cf6b8a10e89280ba73eeef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.809504 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(93487ffefc09be98021c5453fb6a2dc904b9b8a5e9cf6b8a10e89280ba73eeef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.809528 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(93487ffefc09be98021c5453fb6a2dc904b9b8a5e9cf6b8a10e89280ba73eeef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:10:46 crc kubenswrapper[4777]: E1124 17:10:46.809568 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-22cbh_openshift-operators(c0fe98e7-9ad1-426e-a17a-ad471a7f73af)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-22cbh_openshift-operators(c0fe98e7-9ad1-426e-a17a-ad471a7f73af)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(93487ffefc09be98021c5453fb6a2dc904b9b8a5e9cf6b8a10e89280ba73eeef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-22cbh" podUID="c0fe98e7-9ad1-426e-a17a-ad471a7f73af" Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.954046 4777 generic.go:334] "Generic (PLEG): container finished" podID="0f25f5a5-fdf6-4c44-9849-ece6ef97b698" containerID="a4cfcf58dc2c0d11bbc6ad27a962a46067dcfee5214ea4bb53239cc6648b44a6" exitCode=0 Nov 24 17:10:46 crc kubenswrapper[4777]: I1124 17:10:46.954112 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" event={"ID":"0f25f5a5-fdf6-4c44-9849-ece6ef97b698","Type":"ContainerDied","Data":"a4cfcf58dc2c0d11bbc6ad27a962a46067dcfee5214ea4bb53239cc6648b44a6"} Nov 24 17:10:47 crc kubenswrapper[4777]: I1124 17:10:47.257101 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9084c33d-bd72-470c-9f45-3b0b0f29aa19" path="/var/lib/kubelet/pods/9084c33d-bd72-470c-9f45-3b0b0f29aa19/volumes" Nov 24 17:10:47 crc kubenswrapper[4777]: I1124 17:10:47.963426 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" event={"ID":"0f25f5a5-fdf6-4c44-9849-ece6ef97b698","Type":"ContainerStarted","Data":"1c0ba7cf366ba4c3bb2d294e9d575823b8f904d5c239d74ab88c597011430bc1"} Nov 24 17:10:47 crc kubenswrapper[4777]: I1124 17:10:47.963685 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" event={"ID":"0f25f5a5-fdf6-4c44-9849-ece6ef97b698","Type":"ContainerStarted","Data":"8a7464d60c53ea961ff20eab58e32a421ca3a988530f18f07c3b89f2df9118f0"} Nov 24 17:10:47 crc kubenswrapper[4777]: I1124 17:10:47.963696 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" event={"ID":"0f25f5a5-fdf6-4c44-9849-ece6ef97b698","Type":"ContainerStarted","Data":"4703e16660eb32fc6948ed6f49b50264ec3e29d73e4b04bbca7a6c110c209612"} Nov 24 17:10:47 crc kubenswrapper[4777]: I1124 17:10:47.963705 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" event={"ID":"0f25f5a5-fdf6-4c44-9849-ece6ef97b698","Type":"ContainerStarted","Data":"34d2d3051b78dae31b488e8729a1666bea1695043a94372b83ddde3b2089e9ca"} Nov 24 17:10:47 crc kubenswrapper[4777]: I1124 17:10:47.963713 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" event={"ID":"0f25f5a5-fdf6-4c44-9849-ece6ef97b698","Type":"ContainerStarted","Data":"3caad5e6989eeb30869782e20da149882895de9e237c701fac5b70b9dbcb20c1"} Nov 24 17:10:47 crc kubenswrapper[4777]: I1124 17:10:47.963722 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" event={"ID":"0f25f5a5-fdf6-4c44-9849-ece6ef97b698","Type":"ContainerStarted","Data":"5ff9d4df12db7554dafcb5407cb9c144c77a8bf10f118d8dd6a016c9993bbfd4"} Nov 24 17:10:49 crc kubenswrapper[4777]: I1124 17:10:49.975702 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" event={"ID":"0f25f5a5-fdf6-4c44-9849-ece6ef97b698","Type":"ContainerStarted","Data":"93d3ccf5cae42c4fe2f2da63c69fbdcb7194982108e673aaacd3d21b7cb61fa1"} Nov 24 17:10:52 crc kubenswrapper[4777]: I1124 17:10:52.891670 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r"] Nov 24 17:10:52 crc kubenswrapper[4777]: I1124 17:10:52.892459 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:10:52 crc kubenswrapper[4777]: I1124 17:10:52.892825 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:10:52 crc kubenswrapper[4777]: I1124 17:10:52.899332 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq"] Nov 24 17:10:52 crc kubenswrapper[4777]: I1124 17:10:52.899468 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:10:52 crc kubenswrapper[4777]: I1124 17:10:52.899991 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:10:52 crc kubenswrapper[4777]: I1124 17:10:52.918867 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-hdxt8"] Nov 24 17:10:52 crc kubenswrapper[4777]: I1124 17:10:52.918995 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:10:52 crc kubenswrapper[4777]: I1124 17:10:52.919346 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:10:52 crc kubenswrapper[4777]: I1124 17:10:52.937938 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-22cbh"] Nov 24 17:10:52 crc kubenswrapper[4777]: I1124 17:10:52.938134 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:10:52 crc kubenswrapper[4777]: I1124 17:10:52.938721 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:10:52 crc kubenswrapper[4777]: I1124 17:10:52.950785 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx"] Nov 24 17:10:52 crc kubenswrapper[4777]: I1124 17:10:52.950899 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:10:52 crc kubenswrapper[4777]: I1124 17:10:52.951311 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:10:52 crc kubenswrapper[4777]: E1124 17:10:52.975361 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(49e2bc135b595b5a7c580cbec0eea936ab7f444665e8668396888b41129662ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:10:52 crc kubenswrapper[4777]: E1124 17:10:52.975480 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(49e2bc135b595b5a7c580cbec0eea936ab7f444665e8668396888b41129662ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:10:52 crc kubenswrapper[4777]: E1124 17:10:52.975554 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(49e2bc135b595b5a7c580cbec0eea936ab7f444665e8668396888b41129662ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:10:52 crc kubenswrapper[4777]: E1124 17:10:52.975647 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators(dd6addca-3fb6-4204-ae4e-2ed8d1326699)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators(dd6addca-3fb6-4204-ae4e-2ed8d1326699)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(49e2bc135b595b5a7c580cbec0eea936ab7f444665e8668396888b41129662ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" podUID="dd6addca-3fb6-4204-ae4e-2ed8d1326699" Nov 24 17:10:52 crc kubenswrapper[4777]: E1124 17:10:52.986960 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators_67418bcd-f0ba-4526-afc1-8420cb642ccf_0(a33bef2cff21870c51fcdb92930ae18b0d5db74456419d517a71e3fb196433f0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:10:52 crc kubenswrapper[4777]: E1124 17:10:52.987038 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators_67418bcd-f0ba-4526-afc1-8420cb642ccf_0(a33bef2cff21870c51fcdb92930ae18b0d5db74456419d517a71e3fb196433f0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:10:52 crc kubenswrapper[4777]: E1124 17:10:52.987088 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators_67418bcd-f0ba-4526-afc1-8420cb642ccf_0(a33bef2cff21870c51fcdb92930ae18b0d5db74456419d517a71e3fb196433f0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:10:52 crc kubenswrapper[4777]: E1124 17:10:52.987130 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators(67418bcd-f0ba-4526-afc1-8420cb642ccf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators(67418bcd-f0ba-4526-afc1-8420cb642ccf)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators_67418bcd-f0ba-4526-afc1-8420cb642ccf_0(a33bef2cff21870c51fcdb92930ae18b0d5db74456419d517a71e3fb196433f0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" podUID="67418bcd-f0ba-4526-afc1-8420cb642ccf" Nov 24 17:10:52 crc kubenswrapper[4777]: E1124 17:10:52.992929 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-hdxt8_openshift-operators_986d062f-3210-471c-a2db-07e080b5d449_0(551996d581f48f1f5ca8550d6027080b96565a889e65be0308e11b0aefb2ab31): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:10:52 crc kubenswrapper[4777]: E1124 17:10:52.992997 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-hdxt8_openshift-operators_986d062f-3210-471c-a2db-07e080b5d449_0(551996d581f48f1f5ca8550d6027080b96565a889e65be0308e11b0aefb2ab31): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:10:52 crc kubenswrapper[4777]: E1124 17:10:52.993019 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-hdxt8_openshift-operators_986d062f-3210-471c-a2db-07e080b5d449_0(551996d581f48f1f5ca8550d6027080b96565a889e65be0308e11b0aefb2ab31): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:10:52 crc kubenswrapper[4777]: E1124 17:10:52.993059 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-hdxt8_openshift-operators(986d062f-3210-471c-a2db-07e080b5d449)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-hdxt8_openshift-operators(986d062f-3210-471c-a2db-07e080b5d449)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-hdxt8_openshift-operators_986d062f-3210-471c-a2db-07e080b5d449_0(551996d581f48f1f5ca8550d6027080b96565a889e65be0308e11b0aefb2ab31): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" podUID="986d062f-3210-471c-a2db-07e080b5d449" Nov 24 17:10:53 crc kubenswrapper[4777]: I1124 17:10:53.015894 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" event={"ID":"0f25f5a5-fdf6-4c44-9849-ece6ef97b698","Type":"ContainerStarted","Data":"415465ed5ef6ec87a942c73d5aeefbde69ff52707e30204692a517953167b71f"} Nov 24 17:10:53 crc kubenswrapper[4777]: I1124 17:10:53.016745 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:53 crc kubenswrapper[4777]: I1124 17:10:53.016906 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:53 crc kubenswrapper[4777]: I1124 17:10:53.016925 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:53 crc kubenswrapper[4777]: E1124 17:10:53.017023 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(fa486bc85b3d05ff8a82c6dabaaf0b3ece56be2d1b88ed851eeb27fc0be3063f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:10:53 crc kubenswrapper[4777]: E1124 17:10:53.017054 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(fa486bc85b3d05ff8a82c6dabaaf0b3ece56be2d1b88ed851eeb27fc0be3063f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:10:53 crc kubenswrapper[4777]: E1124 17:10:53.017071 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(fa486bc85b3d05ff8a82c6dabaaf0b3ece56be2d1b88ed851eeb27fc0be3063f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:10:53 crc kubenswrapper[4777]: E1124 17:10:53.017104 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-22cbh_openshift-operators(c0fe98e7-9ad1-426e-a17a-ad471a7f73af)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-22cbh_openshift-operators(c0fe98e7-9ad1-426e-a17a-ad471a7f73af)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(fa486bc85b3d05ff8a82c6dabaaf0b3ece56be2d1b88ed851eeb27fc0be3063f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-22cbh" podUID="c0fe98e7-9ad1-426e-a17a-ad471a7f73af" Nov 24 17:10:53 crc kubenswrapper[4777]: E1124 17:10:53.034211 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators_b08a2563-7e34-40f7-b9b0-052824422cff_0(b4aebbad0cfdc7a515e97d4f0c69ec7898fe54ce68ee768364fa04b83a3d276c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:10:53 crc kubenswrapper[4777]: E1124 17:10:53.034270 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators_b08a2563-7e34-40f7-b9b0-052824422cff_0(b4aebbad0cfdc7a515e97d4f0c69ec7898fe54ce68ee768364fa04b83a3d276c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:10:53 crc kubenswrapper[4777]: E1124 17:10:53.034293 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators_b08a2563-7e34-40f7-b9b0-052824422cff_0(b4aebbad0cfdc7a515e97d4f0c69ec7898fe54ce68ee768364fa04b83a3d276c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:10:53 crc kubenswrapper[4777]: E1124 17:10:53.034338 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators(b08a2563-7e34-40f7-b9b0-052824422cff)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators(b08a2563-7e34-40f7-b9b0-052824422cff)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators_b08a2563-7e34-40f7-b9b0-052824422cff_0(b4aebbad0cfdc7a515e97d4f0c69ec7898fe54ce68ee768364fa04b83a3d276c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" podUID="b08a2563-7e34-40f7-b9b0-052824422cff" Nov 24 17:10:53 crc kubenswrapper[4777]: I1124 17:10:53.046525 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:53 crc kubenswrapper[4777]: I1124 17:10:53.056138 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:10:53 crc kubenswrapper[4777]: I1124 17:10:53.070081 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" podStartSLOduration=8.070061591 podStartE2EDuration="8.070061591s" podCreationTimestamp="2025-11-24 17:10:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:10:53.044505216 +0000 UTC m=+611.203700275" watchObservedRunningTime="2025-11-24 17:10:53.070061591 +0000 UTC m=+611.229256640" Nov 24 17:11:00 crc kubenswrapper[4777]: I1124 17:11:00.245007 4777 scope.go:117] "RemoveContainer" containerID="f481bf73ae67258e05eff97b7b2653b611d0cd13cac8da9e0655f728f5bc954f" Nov 24 17:11:00 crc kubenswrapper[4777]: E1124 17:11:00.246891 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-mfvtr_openshift-multus(a5500f09-0e35-442c-87ae-8f280cd2edd1)\"" pod="openshift-multus/multus-mfvtr" podUID="a5500f09-0e35-442c-87ae-8f280cd2edd1" Nov 24 17:11:04 crc kubenswrapper[4777]: I1124 17:11:04.244934 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:11:04 crc kubenswrapper[4777]: I1124 17:11:04.244987 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:11:04 crc kubenswrapper[4777]: I1124 17:11:04.245922 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:11:04 crc kubenswrapper[4777]: I1124 17:11:04.246258 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:11:04 crc kubenswrapper[4777]: I1124 17:11:04.245101 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:11:04 crc kubenswrapper[4777]: I1124 17:11:04.247207 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:11:04 crc kubenswrapper[4777]: E1124 17:11:04.302333 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators_b08a2563-7e34-40f7-b9b0-052824422cff_0(afcc46f0b5050e773f544ceb9a7a561eba367e472c54e113de1edede34c5f370): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:11:04 crc kubenswrapper[4777]: E1124 17:11:04.302428 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators_b08a2563-7e34-40f7-b9b0-052824422cff_0(afcc46f0b5050e773f544ceb9a7a561eba367e472c54e113de1edede34c5f370): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:11:04 crc kubenswrapper[4777]: E1124 17:11:04.302465 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators_b08a2563-7e34-40f7-b9b0-052824422cff_0(afcc46f0b5050e773f544ceb9a7a561eba367e472c54e113de1edede34c5f370): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:11:04 crc kubenswrapper[4777]: E1124 17:11:04.302534 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators(b08a2563-7e34-40f7-b9b0-052824422cff)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators(b08a2563-7e34-40f7-b9b0-052824422cff)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-wsbfx_openshift-operators_b08a2563-7e34-40f7-b9b0-052824422cff_0(afcc46f0b5050e773f544ceb9a7a561eba367e472c54e113de1edede34c5f370): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" podUID="b08a2563-7e34-40f7-b9b0-052824422cff" Nov 24 17:11:04 crc kubenswrapper[4777]: E1124 17:11:04.331933 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(a92bb663046c9d6a162bc7a9b90e142cdf452e309a5ef7da9dc31ca1eda1bc81): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:11:04 crc kubenswrapper[4777]: E1124 17:11:04.332093 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(a92bb663046c9d6a162bc7a9b90e142cdf452e309a5ef7da9dc31ca1eda1bc81): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:11:04 crc kubenswrapper[4777]: E1124 17:11:04.332614 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(a92bb663046c9d6a162bc7a9b90e142cdf452e309a5ef7da9dc31ca1eda1bc81): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:11:04 crc kubenswrapper[4777]: E1124 17:11:04.332678 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-22cbh_openshift-operators(c0fe98e7-9ad1-426e-a17a-ad471a7f73af)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-22cbh_openshift-operators(c0fe98e7-9ad1-426e-a17a-ad471a7f73af)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(a92bb663046c9d6a162bc7a9b90e142cdf452e309a5ef7da9dc31ca1eda1bc81): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-22cbh" podUID="c0fe98e7-9ad1-426e-a17a-ad471a7f73af" Nov 24 17:11:04 crc kubenswrapper[4777]: E1124 17:11:04.338366 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(b96b30ce46d71fc496b29f8f3ae9db2225699c38959d9fdef9086d9065df3814): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:11:04 crc kubenswrapper[4777]: E1124 17:11:04.338453 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(b96b30ce46d71fc496b29f8f3ae9db2225699c38959d9fdef9086d9065df3814): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:11:04 crc kubenswrapper[4777]: E1124 17:11:04.338493 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(b96b30ce46d71fc496b29f8f3ae9db2225699c38959d9fdef9086d9065df3814): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:11:04 crc kubenswrapper[4777]: E1124 17:11:04.338569 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators(dd6addca-3fb6-4204-ae4e-2ed8d1326699)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators(dd6addca-3fb6-4204-ae4e-2ed8d1326699)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(b96b30ce46d71fc496b29f8f3ae9db2225699c38959d9fdef9086d9065df3814): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" podUID="dd6addca-3fb6-4204-ae4e-2ed8d1326699" Nov 24 17:11:07 crc kubenswrapper[4777]: I1124 17:11:07.244824 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:11:07 crc kubenswrapper[4777]: I1124 17:11:07.245540 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:11:07 crc kubenswrapper[4777]: E1124 17:11:07.277130 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-hdxt8_openshift-operators_986d062f-3210-471c-a2db-07e080b5d449_0(29145579de740092db9f3badd6faccb2c4f0445901ad6d25b255fa923c639135): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:11:07 crc kubenswrapper[4777]: E1124 17:11:07.277220 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-hdxt8_openshift-operators_986d062f-3210-471c-a2db-07e080b5d449_0(29145579de740092db9f3badd6faccb2c4f0445901ad6d25b255fa923c639135): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:11:07 crc kubenswrapper[4777]: E1124 17:11:07.277257 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-hdxt8_openshift-operators_986d062f-3210-471c-a2db-07e080b5d449_0(29145579de740092db9f3badd6faccb2c4f0445901ad6d25b255fa923c639135): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:11:07 crc kubenswrapper[4777]: E1124 17:11:07.277328 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-hdxt8_openshift-operators(986d062f-3210-471c-a2db-07e080b5d449)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-hdxt8_openshift-operators(986d062f-3210-471c-a2db-07e080b5d449)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-hdxt8_openshift-operators_986d062f-3210-471c-a2db-07e080b5d449_0(29145579de740092db9f3badd6faccb2c4f0445901ad6d25b255fa923c639135): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" podUID="986d062f-3210-471c-a2db-07e080b5d449" Nov 24 17:11:08 crc kubenswrapper[4777]: I1124 17:11:08.244204 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:11:08 crc kubenswrapper[4777]: I1124 17:11:08.245161 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:11:08 crc kubenswrapper[4777]: E1124 17:11:08.278805 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators_67418bcd-f0ba-4526-afc1-8420cb642ccf_0(9f23653b992776665dbc6d94268c522dc2cec5356c36e5728ec0d77018e01186): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:11:08 crc kubenswrapper[4777]: E1124 17:11:08.278894 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators_67418bcd-f0ba-4526-afc1-8420cb642ccf_0(9f23653b992776665dbc6d94268c522dc2cec5356c36e5728ec0d77018e01186): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:11:08 crc kubenswrapper[4777]: E1124 17:11:08.278934 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators_67418bcd-f0ba-4526-afc1-8420cb642ccf_0(9f23653b992776665dbc6d94268c522dc2cec5356c36e5728ec0d77018e01186): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:11:08 crc kubenswrapper[4777]: E1124 17:11:08.279038 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators(67418bcd-f0ba-4526-afc1-8420cb642ccf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators(67418bcd-f0ba-4526-afc1-8420cb642ccf)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_openshift-operators_67418bcd-f0ba-4526-afc1-8420cb642ccf_0(9f23653b992776665dbc6d94268c522dc2cec5356c36e5728ec0d77018e01186): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" podUID="67418bcd-f0ba-4526-afc1-8420cb642ccf" Nov 24 17:11:14 crc kubenswrapper[4777]: I1124 17:11:14.245529 4777 scope.go:117] "RemoveContainer" containerID="f481bf73ae67258e05eff97b7b2653b611d0cd13cac8da9e0655f728f5bc954f" Nov 24 17:11:15 crc kubenswrapper[4777]: I1124 17:11:15.151684 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mfvtr_a5500f09-0e35-442c-87ae-8f280cd2edd1/kube-multus/2.log" Nov 24 17:11:15 crc kubenswrapper[4777]: I1124 17:11:15.152696 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mfvtr_a5500f09-0e35-442c-87ae-8f280cd2edd1/kube-multus/1.log" Nov 24 17:11:15 crc kubenswrapper[4777]: I1124 17:11:15.152783 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mfvtr" event={"ID":"a5500f09-0e35-442c-87ae-8f280cd2edd1","Type":"ContainerStarted","Data":"2707a332c5f666865e20cc8b96561a1c938a97ae2ee290233da927ce837f8ee9"} Nov 24 17:11:15 crc kubenswrapper[4777]: I1124 17:11:15.244869 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:11:15 crc kubenswrapper[4777]: I1124 17:11:15.245013 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:11:15 crc kubenswrapper[4777]: I1124 17:11:15.245801 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:11:15 crc kubenswrapper[4777]: I1124 17:11:15.245826 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:11:15 crc kubenswrapper[4777]: E1124 17:11:15.291632 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(d1acb8a01e8b7391d83a7f0777e34938b3823bde37d1825ac66667613a6d80f2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:11:15 crc kubenswrapper[4777]: E1124 17:11:15.291727 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(d1acb8a01e8b7391d83a7f0777e34938b3823bde37d1825ac66667613a6d80f2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:11:15 crc kubenswrapper[4777]: E1124 17:11:15.291770 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(d1acb8a01e8b7391d83a7f0777e34938b3823bde37d1825ac66667613a6d80f2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:11:15 crc kubenswrapper[4777]: E1124 17:11:15.291841 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators(dd6addca-3fb6-4204-ae4e-2ed8d1326699)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators(dd6addca-3fb6-4204-ae4e-2ed8d1326699)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_openshift-operators_dd6addca-3fb6-4204-ae4e-2ed8d1326699_0(d1acb8a01e8b7391d83a7f0777e34938b3823bde37d1825ac66667613a6d80f2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" podUID="dd6addca-3fb6-4204-ae4e-2ed8d1326699" Nov 24 17:11:15 crc kubenswrapper[4777]: E1124 17:11:15.307350 4777 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(9241a1c89113506e548b4213b267a40ac157630bacf0a39a4b018e0461d78803): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 17:11:15 crc kubenswrapper[4777]: E1124 17:11:15.307460 4777 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(9241a1c89113506e548b4213b267a40ac157630bacf0a39a4b018e0461d78803): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:11:15 crc kubenswrapper[4777]: E1124 17:11:15.307508 4777 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(9241a1c89113506e548b4213b267a40ac157630bacf0a39a4b018e0461d78803): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:11:15 crc kubenswrapper[4777]: E1124 17:11:15.307608 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-22cbh_openshift-operators(c0fe98e7-9ad1-426e-a17a-ad471a7f73af)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-22cbh_openshift-operators(c0fe98e7-9ad1-426e-a17a-ad471a7f73af)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-22cbh_openshift-operators_c0fe98e7-9ad1-426e-a17a-ad471a7f73af_0(9241a1c89113506e548b4213b267a40ac157630bacf0a39a4b018e0461d78803): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-22cbh" podUID="c0fe98e7-9ad1-426e-a17a-ad471a7f73af" Nov 24 17:11:15 crc kubenswrapper[4777]: I1124 17:11:15.936588 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gxcwh" Nov 24 17:11:18 crc kubenswrapper[4777]: I1124 17:11:18.245102 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:11:18 crc kubenswrapper[4777]: I1124 17:11:18.246140 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" Nov 24 17:11:18 crc kubenswrapper[4777]: I1124 17:11:18.514629 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx"] Nov 24 17:11:18 crc kubenswrapper[4777]: W1124 17:11:18.518878 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb08a2563_7e34_40f7_b9b0_052824422cff.slice/crio-7c3bd2f954743a56724d7ffc0e1596f7bffd1080c7bbbe78639ff261a36b639e WatchSource:0}: Error finding container 7c3bd2f954743a56724d7ffc0e1596f7bffd1080c7bbbe78639ff261a36b639e: Status 404 returned error can't find the container with id 7c3bd2f954743a56724d7ffc0e1596f7bffd1080c7bbbe78639ff261a36b639e Nov 24 17:11:19 crc kubenswrapper[4777]: I1124 17:11:19.192124 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" event={"ID":"b08a2563-7e34-40f7-b9b0-052824422cff","Type":"ContainerStarted","Data":"7c3bd2f954743a56724d7ffc0e1596f7bffd1080c7bbbe78639ff261a36b639e"} Nov 24 17:11:21 crc kubenswrapper[4777]: I1124 17:11:21.244912 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:11:21 crc kubenswrapper[4777]: I1124 17:11:21.245604 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:11:21 crc kubenswrapper[4777]: I1124 17:11:21.490617 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-hdxt8"] Nov 24 17:11:21 crc kubenswrapper[4777]: W1124 17:11:21.510219 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod986d062f_3210_471c_a2db_07e080b5d449.slice/crio-58a69ab667fba4511a7abaeae63857b5d353216da7c41f7bf9b18332227f646e WatchSource:0}: Error finding container 58a69ab667fba4511a7abaeae63857b5d353216da7c41f7bf9b18332227f646e: Status 404 returned error can't find the container with id 58a69ab667fba4511a7abaeae63857b5d353216da7c41f7bf9b18332227f646e Nov 24 17:11:22 crc kubenswrapper[4777]: I1124 17:11:22.210881 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" event={"ID":"986d062f-3210-471c-a2db-07e080b5d449","Type":"ContainerStarted","Data":"58a69ab667fba4511a7abaeae63857b5d353216da7c41f7bf9b18332227f646e"} Nov 24 17:11:23 crc kubenswrapper[4777]: I1124 17:11:23.244609 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:11:23 crc kubenswrapper[4777]: I1124 17:11:23.245107 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" Nov 24 17:11:26 crc kubenswrapper[4777]: I1124 17:11:26.083842 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq"] Nov 24 17:11:26 crc kubenswrapper[4777]: W1124 17:11:26.098981 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67418bcd_f0ba_4526_afc1_8420cb642ccf.slice/crio-7ae7617a20def8cfb0d1038fc395cba43f414e69cd7a7cbcd1b575e35756d36c WatchSource:0}: Error finding container 7ae7617a20def8cfb0d1038fc395cba43f414e69cd7a7cbcd1b575e35756d36c: Status 404 returned error can't find the container with id 7ae7617a20def8cfb0d1038fc395cba43f414e69cd7a7cbcd1b575e35756d36c Nov 24 17:11:26 crc kubenswrapper[4777]: I1124 17:11:26.238854 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" event={"ID":"67418bcd-f0ba-4526-afc1-8420cb642ccf","Type":"ContainerStarted","Data":"7ae7617a20def8cfb0d1038fc395cba43f414e69cd7a7cbcd1b575e35756d36c"} Nov 24 17:11:26 crc kubenswrapper[4777]: I1124 17:11:26.240446 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" event={"ID":"b08a2563-7e34-40f7-b9b0-052824422cff","Type":"ContainerStarted","Data":"2e169e24efdde9810c95d23df391bbffa4be99ca10bcd78841487acd5de42b13"} Nov 24 17:11:26 crc kubenswrapper[4777]: I1124 17:11:26.271687 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-wsbfx" podStartSLOduration=32.841894452 podStartE2EDuration="40.271657624s" podCreationTimestamp="2025-11-24 17:10:46 +0000 UTC" firstStartedPulling="2025-11-24 17:11:18.521603177 +0000 UTC m=+636.680798236" lastFinishedPulling="2025-11-24 17:11:25.951366359 +0000 UTC m=+644.110561408" observedRunningTime="2025-11-24 17:11:26.26505371 +0000 UTC m=+644.424248819" watchObservedRunningTime="2025-11-24 17:11:26.271657624 +0000 UTC m=+644.430852703" Nov 24 17:11:28 crc kubenswrapper[4777]: I1124 17:11:28.244418 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:11:28 crc kubenswrapper[4777]: I1124 17:11:28.245985 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:11:29 crc kubenswrapper[4777]: I1124 17:11:29.248623 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:11:29 crc kubenswrapper[4777]: I1124 17:11:29.249105 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" Nov 24 17:11:30 crc kubenswrapper[4777]: I1124 17:11:30.572236 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r"] Nov 24 17:11:30 crc kubenswrapper[4777]: W1124 17:11:30.579549 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd6addca_3fb6_4204_ae4e_2ed8d1326699.slice/crio-f2f00743136c1e849b6d23aa356609c7265e2bef992c1e6dbe0e706b157fc26d WatchSource:0}: Error finding container f2f00743136c1e849b6d23aa356609c7265e2bef992c1e6dbe0e706b157fc26d: Status 404 returned error can't find the container with id f2f00743136c1e849b6d23aa356609c7265e2bef992c1e6dbe0e706b157fc26d Nov 24 17:11:30 crc kubenswrapper[4777]: I1124 17:11:30.615363 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-22cbh"] Nov 24 17:11:30 crc kubenswrapper[4777]: W1124 17:11:30.620401 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0fe98e7_9ad1_426e_a17a_ad471a7f73af.slice/crio-e5b966acb24b48da83acd9592255781cd49db9f7a0b03107bc794b5e50daec01 WatchSource:0}: Error finding container e5b966acb24b48da83acd9592255781cd49db9f7a0b03107bc794b5e50daec01: Status 404 returned error can't find the container with id e5b966acb24b48da83acd9592255781cd49db9f7a0b03107bc794b5e50daec01 Nov 24 17:11:31 crc kubenswrapper[4777]: I1124 17:11:31.278411 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" event={"ID":"dd6addca-3fb6-4204-ae4e-2ed8d1326699","Type":"ContainerStarted","Data":"f2f00743136c1e849b6d23aa356609c7265e2bef992c1e6dbe0e706b157fc26d"} Nov 24 17:11:31 crc kubenswrapper[4777]: I1124 17:11:31.280074 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" event={"ID":"986d062f-3210-471c-a2db-07e080b5d449","Type":"ContainerStarted","Data":"439c0c9d8e32d91d52c397cf8cd866136fa5b8e777d9df831d915b825f72d5ac"} Nov 24 17:11:31 crc kubenswrapper[4777]: I1124 17:11:31.282172 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" event={"ID":"67418bcd-f0ba-4526-afc1-8420cb642ccf","Type":"ContainerStarted","Data":"ebd208e50a9dd930d425ed8e30135c637125bb9937fea8e396d1e8ba4e1e1411"} Nov 24 17:11:31 crc kubenswrapper[4777]: I1124 17:11:31.283113 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-22cbh" event={"ID":"c0fe98e7-9ad1-426e-a17a-ad471a7f73af","Type":"ContainerStarted","Data":"e5b966acb24b48da83acd9592255781cd49db9f7a0b03107bc794b5e50daec01"} Nov 24 17:11:34 crc kubenswrapper[4777]: I1124 17:11:34.302635 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:11:34 crc kubenswrapper[4777]: I1124 17:11:34.303955 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" Nov 24 17:11:34 crc kubenswrapper[4777]: I1124 17:11:34.317069 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq" podStartSLOduration=44.094050428 podStartE2EDuration="48.317054659s" podCreationTimestamp="2025-11-24 17:10:46 +0000 UTC" firstStartedPulling="2025-11-24 17:11:26.109998852 +0000 UTC m=+644.269193901" lastFinishedPulling="2025-11-24 17:11:30.333003053 +0000 UTC m=+648.492198132" observedRunningTime="2025-11-24 17:11:34.31529902 +0000 UTC m=+652.474494059" watchObservedRunningTime="2025-11-24 17:11:34.317054659 +0000 UTC m=+652.476249708" Nov 24 17:11:34 crc kubenswrapper[4777]: I1124 17:11:34.340162 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-hdxt8" podStartSLOduration=39.522805463 podStartE2EDuration="48.340147381s" podCreationTimestamp="2025-11-24 17:10:46 +0000 UTC" firstStartedPulling="2025-11-24 17:11:21.51413572 +0000 UTC m=+639.673330789" lastFinishedPulling="2025-11-24 17:11:30.331477608 +0000 UTC m=+648.490672707" observedRunningTime="2025-11-24 17:11:34.336723836 +0000 UTC m=+652.495918885" watchObservedRunningTime="2025-11-24 17:11:34.340147381 +0000 UTC m=+652.499342430" Nov 24 17:11:35 crc kubenswrapper[4777]: I1124 17:11:35.310844 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" event={"ID":"dd6addca-3fb6-4204-ae4e-2ed8d1326699","Type":"ContainerStarted","Data":"9f7f208144148ec582c7947dd5f7136ab1ce15fcebc57e6c7b3baf2c32ef3ae5"} Nov 24 17:11:35 crc kubenswrapper[4777]: I1124 17:11:35.352569 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r" podStartSLOduration=49.352532071 podStartE2EDuration="49.352532071s" podCreationTimestamp="2025-11-24 17:10:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:11:35.343292655 +0000 UTC m=+653.502487774" watchObservedRunningTime="2025-11-24 17:11:35.352532071 +0000 UTC m=+653.511727160" Nov 24 17:11:38 crc kubenswrapper[4777]: I1124 17:11:38.337609 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-22cbh" event={"ID":"c0fe98e7-9ad1-426e-a17a-ad471a7f73af","Type":"ContainerStarted","Data":"2054abb8d1760b73db3888ade087a77655c0c35c25f34b741f921d2a3f186638"} Nov 24 17:11:39 crc kubenswrapper[4777]: I1124 17:11:39.342118 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:11:39 crc kubenswrapper[4777]: I1124 17:11:39.367676 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-22cbh" podStartSLOduration=45.881674353 podStartE2EDuration="53.367646817s" podCreationTimestamp="2025-11-24 17:10:46 +0000 UTC" firstStartedPulling="2025-11-24 17:11:30.624159161 +0000 UTC m=+648.783354230" lastFinishedPulling="2025-11-24 17:11:38.110131595 +0000 UTC m=+656.269326694" observedRunningTime="2025-11-24 17:11:39.364995033 +0000 UTC m=+657.524190082" watchObservedRunningTime="2025-11-24 17:11:39.367646817 +0000 UTC m=+657.526841906" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.010509 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-596z9"] Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.011441 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-596z9" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.014745 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.015043 4777 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-c7jbd" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.015271 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.018399 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-l75dg"] Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.019319 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-l75dg" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.022844 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-596z9"] Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.024258 4777 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-6l64k" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.032315 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-l75dg"] Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.043387 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-9f892"] Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.044158 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-9f892" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.048404 4777 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-q59h4" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.067503 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-9f892"] Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.076434 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmllc\" (UniqueName: \"kubernetes.io/projected/0a75da4d-a94d-4639-ad88-0b415b87fe24-kube-api-access-mmllc\") pod \"cert-manager-5b446d88c5-l75dg\" (UID: \"0a75da4d-a94d-4639-ad88-0b415b87fe24\") " pod="cert-manager/cert-manager-5b446d88c5-l75dg" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.076504 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcw68\" (UniqueName: \"kubernetes.io/projected/cee47616-4825-45f3-8232-4f46a66c954a-kube-api-access-vcw68\") pod \"cert-manager-cainjector-7f985d654d-596z9\" (UID: \"cee47616-4825-45f3-8232-4f46a66c954a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-596z9" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.076539 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-479bp\" (UniqueName: \"kubernetes.io/projected/c32a0d11-9b78-4ce6-a4db-d9d07c9f3313-kube-api-access-479bp\") pod \"cert-manager-webhook-5655c58dd6-9f892\" (UID: \"c32a0d11-9b78-4ce6-a4db-d9d07c9f3313\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-9f892" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.177484 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-479bp\" (UniqueName: \"kubernetes.io/projected/c32a0d11-9b78-4ce6-a4db-d9d07c9f3313-kube-api-access-479bp\") pod \"cert-manager-webhook-5655c58dd6-9f892\" (UID: \"c32a0d11-9b78-4ce6-a4db-d9d07c9f3313\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-9f892" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.177544 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmllc\" (UniqueName: \"kubernetes.io/projected/0a75da4d-a94d-4639-ad88-0b415b87fe24-kube-api-access-mmllc\") pod \"cert-manager-5b446d88c5-l75dg\" (UID: \"0a75da4d-a94d-4639-ad88-0b415b87fe24\") " pod="cert-manager/cert-manager-5b446d88c5-l75dg" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.177595 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcw68\" (UniqueName: \"kubernetes.io/projected/cee47616-4825-45f3-8232-4f46a66c954a-kube-api-access-vcw68\") pod \"cert-manager-cainjector-7f985d654d-596z9\" (UID: \"cee47616-4825-45f3-8232-4f46a66c954a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-596z9" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.194141 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.201554 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.211194 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-479bp\" (UniqueName: \"kubernetes.io/projected/c32a0d11-9b78-4ce6-a4db-d9d07c9f3313-kube-api-access-479bp\") pod \"cert-manager-webhook-5655c58dd6-9f892\" (UID: \"c32a0d11-9b78-4ce6-a4db-d9d07c9f3313\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-9f892" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.211696 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcw68\" (UniqueName: \"kubernetes.io/projected/cee47616-4825-45f3-8232-4f46a66c954a-kube-api-access-vcw68\") pod \"cert-manager-cainjector-7f985d654d-596z9\" (UID: \"cee47616-4825-45f3-8232-4f46a66c954a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-596z9" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.213232 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmllc\" (UniqueName: \"kubernetes.io/projected/0a75da4d-a94d-4639-ad88-0b415b87fe24-kube-api-access-mmllc\") pod \"cert-manager-5b446d88c5-l75dg\" (UID: \"0a75da4d-a94d-4639-ad88-0b415b87fe24\") " pod="cert-manager/cert-manager-5b446d88c5-l75dg" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.332656 4777 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-c7jbd" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.340647 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-596z9" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.343014 4777 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-6l64k" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.350423 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-l75dg" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.363838 4777 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-q59h4" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.372695 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-9f892" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.471693 4777 scope.go:117] "RemoveContainer" containerID="899132c850ac1d7af915f5cb557c347ff8caec6f210b43743788369f5852c7e8" Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.594265 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-596z9"] Nov 24 17:11:43 crc kubenswrapper[4777]: W1124 17:11:43.605300 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcee47616_4825_45f3_8232_4f46a66c954a.slice/crio-02200cf394958f6ca8f75fa635b51215c3d6bf7f67ec71e8824b40dcc1a907bb WatchSource:0}: Error finding container 02200cf394958f6ca8f75fa635b51215c3d6bf7f67ec71e8824b40dcc1a907bb: Status 404 returned error can't find the container with id 02200cf394958f6ca8f75fa635b51215c3d6bf7f67ec71e8824b40dcc1a907bb Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.875337 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-9f892"] Nov 24 17:11:43 crc kubenswrapper[4777]: W1124 17:11:43.876642 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc32a0d11_9b78_4ce6_a4db_d9d07c9f3313.slice/crio-04656c34010dca8a29135c1030f72c22fcbcc46447df2d16a1a0f337e23fa612 WatchSource:0}: Error finding container 04656c34010dca8a29135c1030f72c22fcbcc46447df2d16a1a0f337e23fa612: Status 404 returned error can't find the container with id 04656c34010dca8a29135c1030f72c22fcbcc46447df2d16a1a0f337e23fa612 Nov 24 17:11:43 crc kubenswrapper[4777]: W1124 17:11:43.879570 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a75da4d_a94d_4639_ad88_0b415b87fe24.slice/crio-afc294ccfc266503bb1a931eb7e2ef81951be58fcb97db65813ec276c92e193d WatchSource:0}: Error finding container afc294ccfc266503bb1a931eb7e2ef81951be58fcb97db65813ec276c92e193d: Status 404 returned error can't find the container with id afc294ccfc266503bb1a931eb7e2ef81951be58fcb97db65813ec276c92e193d Nov 24 17:11:43 crc kubenswrapper[4777]: I1124 17:11:43.883291 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-l75dg"] Nov 24 17:11:44 crc kubenswrapper[4777]: I1124 17:11:44.393929 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-l75dg" event={"ID":"0a75da4d-a94d-4639-ad88-0b415b87fe24","Type":"ContainerStarted","Data":"afc294ccfc266503bb1a931eb7e2ef81951be58fcb97db65813ec276c92e193d"} Nov 24 17:11:44 crc kubenswrapper[4777]: I1124 17:11:44.408052 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-596z9" event={"ID":"cee47616-4825-45f3-8232-4f46a66c954a","Type":"ContainerStarted","Data":"02200cf394958f6ca8f75fa635b51215c3d6bf7f67ec71e8824b40dcc1a907bb"} Nov 24 17:11:44 crc kubenswrapper[4777]: I1124 17:11:44.409501 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-9f892" event={"ID":"c32a0d11-9b78-4ce6-a4db-d9d07c9f3313","Type":"ContainerStarted","Data":"04656c34010dca8a29135c1030f72c22fcbcc46447df2d16a1a0f337e23fa612"} Nov 24 17:11:44 crc kubenswrapper[4777]: I1124 17:11:44.411734 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mfvtr_a5500f09-0e35-442c-87ae-8f280cd2edd1/kube-multus/2.log" Nov 24 17:11:46 crc kubenswrapper[4777]: I1124 17:11:46.795201 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-22cbh" Nov 24 17:11:47 crc kubenswrapper[4777]: I1124 17:11:47.427035 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-596z9" event={"ID":"cee47616-4825-45f3-8232-4f46a66c954a","Type":"ContainerStarted","Data":"962e0222690664f848349def585fa5ce03497e2d3704945995dbdd1e12fdd01c"} Nov 24 17:11:47 crc kubenswrapper[4777]: I1124 17:11:47.428448 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-9f892" event={"ID":"c32a0d11-9b78-4ce6-a4db-d9d07c9f3313","Type":"ContainerStarted","Data":"a6ac74bc65fc8208ad71b757bd59e62c4c351a48cee5ea7c2abde3d05780a784"} Nov 24 17:11:47 crc kubenswrapper[4777]: I1124 17:11:47.428568 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-9f892" Nov 24 17:11:47 crc kubenswrapper[4777]: I1124 17:11:47.429704 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-l75dg" event={"ID":"0a75da4d-a94d-4639-ad88-0b415b87fe24","Type":"ContainerStarted","Data":"2aee33c93f380ae49a05af765a0dadb2e9238badd0e1d1ad1095da829017dab8"} Nov 24 17:11:47 crc kubenswrapper[4777]: I1124 17:11:47.440253 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-596z9" podStartSLOduration=1.944245534 podStartE2EDuration="5.440236015s" podCreationTimestamp="2025-11-24 17:11:42 +0000 UTC" firstStartedPulling="2025-11-24 17:11:43.607287501 +0000 UTC m=+661.766482570" lastFinishedPulling="2025-11-24 17:11:47.103278002 +0000 UTC m=+665.262473051" observedRunningTime="2025-11-24 17:11:47.439040852 +0000 UTC m=+665.598235891" watchObservedRunningTime="2025-11-24 17:11:47.440236015 +0000 UTC m=+665.599431064" Nov 24 17:11:47 crc kubenswrapper[4777]: I1124 17:11:47.452935 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-9f892" podStartSLOduration=1.177127595 podStartE2EDuration="4.452919407s" podCreationTimestamp="2025-11-24 17:11:43 +0000 UTC" firstStartedPulling="2025-11-24 17:11:43.879644879 +0000 UTC m=+662.038839928" lastFinishedPulling="2025-11-24 17:11:47.155436671 +0000 UTC m=+665.314631740" observedRunningTime="2025-11-24 17:11:47.452162286 +0000 UTC m=+665.611357335" watchObservedRunningTime="2025-11-24 17:11:47.452919407 +0000 UTC m=+665.612114456" Nov 24 17:11:47 crc kubenswrapper[4777]: I1124 17:11:47.470160 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-l75dg" podStartSLOduration=2.25536102 podStartE2EDuration="5.470141206s" podCreationTimestamp="2025-11-24 17:11:42 +0000 UTC" firstStartedPulling="2025-11-24 17:11:43.881709907 +0000 UTC m=+662.040904956" lastFinishedPulling="2025-11-24 17:11:47.096490093 +0000 UTC m=+665.255685142" observedRunningTime="2025-11-24 17:11:47.467949905 +0000 UTC m=+665.627144954" watchObservedRunningTime="2025-11-24 17:11:47.470141206 +0000 UTC m=+665.629336255" Nov 24 17:11:53 crc kubenswrapper[4777]: I1124 17:11:53.378169 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-9f892" Nov 24 17:12:14 crc kubenswrapper[4777]: I1124 17:12:14.270343 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:12:14 crc kubenswrapper[4777]: I1124 17:12:14.271138 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:12:18 crc kubenswrapper[4777]: I1124 17:12:18.649816 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd"] Nov 24 17:12:18 crc kubenswrapper[4777]: I1124 17:12:18.651004 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" Nov 24 17:12:18 crc kubenswrapper[4777]: I1124 17:12:18.652469 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 17:12:18 crc kubenswrapper[4777]: I1124 17:12:18.662395 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd"] Nov 24 17:12:18 crc kubenswrapper[4777]: I1124 17:12:18.747501 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/edece1c5-1d2f-4a61-9591-7313cd41d58d-util\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd\" (UID: \"edece1c5-1d2f-4a61-9591-7313cd41d58d\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" Nov 24 17:12:18 crc kubenswrapper[4777]: I1124 17:12:18.747752 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/edece1c5-1d2f-4a61-9591-7313cd41d58d-bundle\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd\" (UID: \"edece1c5-1d2f-4a61-9591-7313cd41d58d\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" Nov 24 17:12:18 crc kubenswrapper[4777]: I1124 17:12:18.747853 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm7zp\" (UniqueName: \"kubernetes.io/projected/edece1c5-1d2f-4a61-9591-7313cd41d58d-kube-api-access-hm7zp\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd\" (UID: \"edece1c5-1d2f-4a61-9591-7313cd41d58d\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" Nov 24 17:12:18 crc kubenswrapper[4777]: I1124 17:12:18.849442 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/edece1c5-1d2f-4a61-9591-7313cd41d58d-util\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd\" (UID: \"edece1c5-1d2f-4a61-9591-7313cd41d58d\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" Nov 24 17:12:18 crc kubenswrapper[4777]: I1124 17:12:18.849522 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/edece1c5-1d2f-4a61-9591-7313cd41d58d-bundle\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd\" (UID: \"edece1c5-1d2f-4a61-9591-7313cd41d58d\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" Nov 24 17:12:18 crc kubenswrapper[4777]: I1124 17:12:18.849552 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm7zp\" (UniqueName: \"kubernetes.io/projected/edece1c5-1d2f-4a61-9591-7313cd41d58d-kube-api-access-hm7zp\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd\" (UID: \"edece1c5-1d2f-4a61-9591-7313cd41d58d\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" Nov 24 17:12:18 crc kubenswrapper[4777]: I1124 17:12:18.850246 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/edece1c5-1d2f-4a61-9591-7313cd41d58d-bundle\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd\" (UID: \"edece1c5-1d2f-4a61-9591-7313cd41d58d\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" Nov 24 17:12:18 crc kubenswrapper[4777]: I1124 17:12:18.850305 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/edece1c5-1d2f-4a61-9591-7313cd41d58d-util\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd\" (UID: \"edece1c5-1d2f-4a61-9591-7313cd41d58d\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" Nov 24 17:12:18 crc kubenswrapper[4777]: I1124 17:12:18.886674 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm7zp\" (UniqueName: \"kubernetes.io/projected/edece1c5-1d2f-4a61-9591-7313cd41d58d-kube-api-access-hm7zp\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd\" (UID: \"edece1c5-1d2f-4a61-9591-7313cd41d58d\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" Nov 24 17:12:18 crc kubenswrapper[4777]: I1124 17:12:18.966020 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" Nov 24 17:12:19 crc kubenswrapper[4777]: I1124 17:12:19.456180 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd"] Nov 24 17:12:19 crc kubenswrapper[4777]: I1124 17:12:19.652901 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" event={"ID":"edece1c5-1d2f-4a61-9591-7313cd41d58d","Type":"ContainerStarted","Data":"8e9bc832eb1f02ecc477118ca4c5b78165504be96546d0f46d4b214d9bbe936b"} Nov 24 17:12:19 crc kubenswrapper[4777]: I1124 17:12:19.653222 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" event={"ID":"edece1c5-1d2f-4a61-9591-7313cd41d58d","Type":"ContainerStarted","Data":"c6539b90a881908ca92d2ea2019b54c45cf210a722157746532c93bb7443192d"} Nov 24 17:12:20 crc kubenswrapper[4777]: I1124 17:12:20.663756 4777 generic.go:334] "Generic (PLEG): container finished" podID="edece1c5-1d2f-4a61-9591-7313cd41d58d" containerID="8e9bc832eb1f02ecc477118ca4c5b78165504be96546d0f46d4b214d9bbe936b" exitCode=0 Nov 24 17:12:20 crc kubenswrapper[4777]: I1124 17:12:20.663795 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" event={"ID":"edece1c5-1d2f-4a61-9591-7313cd41d58d","Type":"ContainerDied","Data":"8e9bc832eb1f02ecc477118ca4c5b78165504be96546d0f46d4b214d9bbe936b"} Nov 24 17:12:21 crc kubenswrapper[4777]: I1124 17:12:21.959765 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Nov 24 17:12:21 crc kubenswrapper[4777]: I1124 17:12:21.962446 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Nov 24 17:12:21 crc kubenswrapper[4777]: I1124 17:12:21.964733 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Nov 24 17:12:21 crc kubenswrapper[4777]: I1124 17:12:21.965403 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Nov 24 17:12:21 crc kubenswrapper[4777]: I1124 17:12:21.976722 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Nov 24 17:12:22 crc kubenswrapper[4777]: I1124 17:12:22.085152 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8ee8149e-68ef-4cae-a86c-ceeb0713b2c3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8ee8149e-68ef-4cae-a86c-ceeb0713b2c3\") pod \"minio\" (UID: \"df657a28-5ab4-47f8-814a-6ec63b5d3ba8\") " pod="minio-dev/minio" Nov 24 17:12:22 crc kubenswrapper[4777]: I1124 17:12:22.085224 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrgkx\" (UniqueName: \"kubernetes.io/projected/df657a28-5ab4-47f8-814a-6ec63b5d3ba8-kube-api-access-qrgkx\") pod \"minio\" (UID: \"df657a28-5ab4-47f8-814a-6ec63b5d3ba8\") " pod="minio-dev/minio" Nov 24 17:12:22 crc kubenswrapper[4777]: I1124 17:12:22.186575 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8ee8149e-68ef-4cae-a86c-ceeb0713b2c3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8ee8149e-68ef-4cae-a86c-ceeb0713b2c3\") pod \"minio\" (UID: \"df657a28-5ab4-47f8-814a-6ec63b5d3ba8\") " pod="minio-dev/minio" Nov 24 17:12:22 crc kubenswrapper[4777]: I1124 17:12:22.186646 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrgkx\" (UniqueName: \"kubernetes.io/projected/df657a28-5ab4-47f8-814a-6ec63b5d3ba8-kube-api-access-qrgkx\") pod \"minio\" (UID: \"df657a28-5ab4-47f8-814a-6ec63b5d3ba8\") " pod="minio-dev/minio" Nov 24 17:12:22 crc kubenswrapper[4777]: I1124 17:12:22.190550 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:12:22 crc kubenswrapper[4777]: I1124 17:12:22.190610 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8ee8149e-68ef-4cae-a86c-ceeb0713b2c3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8ee8149e-68ef-4cae-a86c-ceeb0713b2c3\") pod \"minio\" (UID: \"df657a28-5ab4-47f8-814a-6ec63b5d3ba8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/760a6ac93b0ad2b669162f279878793de1c481d82e3c4ec1bd8e40822e9a32c8/globalmount\"" pod="minio-dev/minio" Nov 24 17:12:22 crc kubenswrapper[4777]: I1124 17:12:22.226142 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8ee8149e-68ef-4cae-a86c-ceeb0713b2c3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8ee8149e-68ef-4cae-a86c-ceeb0713b2c3\") pod \"minio\" (UID: \"df657a28-5ab4-47f8-814a-6ec63b5d3ba8\") " pod="minio-dev/minio" Nov 24 17:12:22 crc kubenswrapper[4777]: I1124 17:12:22.226459 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrgkx\" (UniqueName: \"kubernetes.io/projected/df657a28-5ab4-47f8-814a-6ec63b5d3ba8-kube-api-access-qrgkx\") pod \"minio\" (UID: \"df657a28-5ab4-47f8-814a-6ec63b5d3ba8\") " pod="minio-dev/minio" Nov 24 17:12:22 crc kubenswrapper[4777]: I1124 17:12:22.286647 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Nov 24 17:12:22 crc kubenswrapper[4777]: I1124 17:12:22.580540 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Nov 24 17:12:22 crc kubenswrapper[4777]: I1124 17:12:22.676738 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"df657a28-5ab4-47f8-814a-6ec63b5d3ba8","Type":"ContainerStarted","Data":"822d9de68208cdf9fc69764687f8a0b99f176c2aa933f219b2c651a82d1a115f"} Nov 24 17:12:23 crc kubenswrapper[4777]: I1124 17:12:23.685333 4777 generic.go:334] "Generic (PLEG): container finished" podID="edece1c5-1d2f-4a61-9591-7313cd41d58d" containerID="80d432f3ccecfde11def438a2b90165ecf15351c357b15e5207f78bbfc69c9c9" exitCode=0 Nov 24 17:12:23 crc kubenswrapper[4777]: I1124 17:12:23.685365 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" event={"ID":"edece1c5-1d2f-4a61-9591-7313cd41d58d","Type":"ContainerDied","Data":"80d432f3ccecfde11def438a2b90165ecf15351c357b15e5207f78bbfc69c9c9"} Nov 24 17:12:24 crc kubenswrapper[4777]: I1124 17:12:24.693329 4777 generic.go:334] "Generic (PLEG): container finished" podID="edece1c5-1d2f-4a61-9591-7313cd41d58d" containerID="f1e0dce6eef6104f10f80a5a9eaa49a40d97d62c64c18a7ec71fa87aa0c46773" exitCode=0 Nov 24 17:12:24 crc kubenswrapper[4777]: I1124 17:12:24.693621 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" event={"ID":"edece1c5-1d2f-4a61-9591-7313cd41d58d","Type":"ContainerDied","Data":"f1e0dce6eef6104f10f80a5a9eaa49a40d97d62c64c18a7ec71fa87aa0c46773"} Nov 24 17:12:26 crc kubenswrapper[4777]: I1124 17:12:26.015744 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" Nov 24 17:12:26 crc kubenswrapper[4777]: I1124 17:12:26.143588 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm7zp\" (UniqueName: \"kubernetes.io/projected/edece1c5-1d2f-4a61-9591-7313cd41d58d-kube-api-access-hm7zp\") pod \"edece1c5-1d2f-4a61-9591-7313cd41d58d\" (UID: \"edece1c5-1d2f-4a61-9591-7313cd41d58d\") " Nov 24 17:12:26 crc kubenswrapper[4777]: I1124 17:12:26.143692 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/edece1c5-1d2f-4a61-9591-7313cd41d58d-bundle\") pod \"edece1c5-1d2f-4a61-9591-7313cd41d58d\" (UID: \"edece1c5-1d2f-4a61-9591-7313cd41d58d\") " Nov 24 17:12:26 crc kubenswrapper[4777]: I1124 17:12:26.143847 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/edece1c5-1d2f-4a61-9591-7313cd41d58d-util\") pod \"edece1c5-1d2f-4a61-9591-7313cd41d58d\" (UID: \"edece1c5-1d2f-4a61-9591-7313cd41d58d\") " Nov 24 17:12:26 crc kubenswrapper[4777]: I1124 17:12:26.145572 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edece1c5-1d2f-4a61-9591-7313cd41d58d-bundle" (OuterVolumeSpecName: "bundle") pod "edece1c5-1d2f-4a61-9591-7313cd41d58d" (UID: "edece1c5-1d2f-4a61-9591-7313cd41d58d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:12:26 crc kubenswrapper[4777]: I1124 17:12:26.154697 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edece1c5-1d2f-4a61-9591-7313cd41d58d-kube-api-access-hm7zp" (OuterVolumeSpecName: "kube-api-access-hm7zp") pod "edece1c5-1d2f-4a61-9591-7313cd41d58d" (UID: "edece1c5-1d2f-4a61-9591-7313cd41d58d"). InnerVolumeSpecName "kube-api-access-hm7zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:12:26 crc kubenswrapper[4777]: I1124 17:12:26.184569 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edece1c5-1d2f-4a61-9591-7313cd41d58d-util" (OuterVolumeSpecName: "util") pod "edece1c5-1d2f-4a61-9591-7313cd41d58d" (UID: "edece1c5-1d2f-4a61-9591-7313cd41d58d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:12:26 crc kubenswrapper[4777]: I1124 17:12:26.245646 4777 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/edece1c5-1d2f-4a61-9591-7313cd41d58d-util\") on node \"crc\" DevicePath \"\"" Nov 24 17:12:26 crc kubenswrapper[4777]: I1124 17:12:26.245688 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hm7zp\" (UniqueName: \"kubernetes.io/projected/edece1c5-1d2f-4a61-9591-7313cd41d58d-kube-api-access-hm7zp\") on node \"crc\" DevicePath \"\"" Nov 24 17:12:26 crc kubenswrapper[4777]: I1124 17:12:26.245710 4777 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/edece1c5-1d2f-4a61-9591-7313cd41d58d-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:12:26 crc kubenswrapper[4777]: I1124 17:12:26.713550 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" event={"ID":"edece1c5-1d2f-4a61-9591-7313cd41d58d","Type":"ContainerDied","Data":"c6539b90a881908ca92d2ea2019b54c45cf210a722157746532c93bb7443192d"} Nov 24 17:12:26 crc kubenswrapper[4777]: I1124 17:12:26.713612 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6539b90a881908ca92d2ea2019b54c45cf210a722157746532c93bb7443192d" Nov 24 17:12:26 crc kubenswrapper[4777]: I1124 17:12:26.713609 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd" Nov 24 17:12:26 crc kubenswrapper[4777]: I1124 17:12:26.716418 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"df657a28-5ab4-47f8-814a-6ec63b5d3ba8","Type":"ContainerStarted","Data":"e4b881d795b357ac3178246c447faabd530072cc9ec4c7ad1f99cf8a0a8eb10a"} Nov 24 17:12:26 crc kubenswrapper[4777]: I1124 17:12:26.737725 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.454460138 podStartE2EDuration="7.737698917s" podCreationTimestamp="2025-11-24 17:12:19 +0000 UTC" firstStartedPulling="2025-11-24 17:12:22.607458613 +0000 UTC m=+700.766653662" lastFinishedPulling="2025-11-24 17:12:25.890697382 +0000 UTC m=+704.049892441" observedRunningTime="2025-11-24 17:12:26.73418275 +0000 UTC m=+704.893377829" watchObservedRunningTime="2025-11-24 17:12:26.737698917 +0000 UTC m=+704.896894006" Nov 24 17:12:31 crc kubenswrapper[4777]: I1124 17:12:31.989185 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c"] Nov 24 17:12:31 crc kubenswrapper[4777]: E1124 17:12:31.989766 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edece1c5-1d2f-4a61-9591-7313cd41d58d" containerName="util" Nov 24 17:12:31 crc kubenswrapper[4777]: I1124 17:12:31.989777 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="edece1c5-1d2f-4a61-9591-7313cd41d58d" containerName="util" Nov 24 17:12:31 crc kubenswrapper[4777]: E1124 17:12:31.989785 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edece1c5-1d2f-4a61-9591-7313cd41d58d" containerName="pull" Nov 24 17:12:31 crc kubenswrapper[4777]: I1124 17:12:31.989790 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="edece1c5-1d2f-4a61-9591-7313cd41d58d" containerName="pull" Nov 24 17:12:31 crc kubenswrapper[4777]: E1124 17:12:31.989801 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edece1c5-1d2f-4a61-9591-7313cd41d58d" containerName="extract" Nov 24 17:12:31 crc kubenswrapper[4777]: I1124 17:12:31.989807 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="edece1c5-1d2f-4a61-9591-7313cd41d58d" containerName="extract" Nov 24 17:12:31 crc kubenswrapper[4777]: I1124 17:12:31.989895 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="edece1c5-1d2f-4a61-9591-7313cd41d58d" containerName="extract" Nov 24 17:12:31 crc kubenswrapper[4777]: I1124 17:12:31.990597 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" Nov 24 17:12:31 crc kubenswrapper[4777]: I1124 17:12:31.996864 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c"] Nov 24 17:12:31 crc kubenswrapper[4777]: I1124 17:12:31.998241 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 17:12:32 crc kubenswrapper[4777]: I1124 17:12:32.085342 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7knx\" (UniqueName: \"kubernetes.io/projected/1ffbfe00-f54c-446d-bc90-d18f24f51f82-kube-api-access-n7knx\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c\" (UID: \"1ffbfe00-f54c-446d-bc90-d18f24f51f82\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" Nov 24 17:12:32 crc kubenswrapper[4777]: I1124 17:12:32.085681 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ffbfe00-f54c-446d-bc90-d18f24f51f82-bundle\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c\" (UID: \"1ffbfe00-f54c-446d-bc90-d18f24f51f82\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" Nov 24 17:12:32 crc kubenswrapper[4777]: I1124 17:12:32.086072 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ffbfe00-f54c-446d-bc90-d18f24f51f82-util\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c\" (UID: \"1ffbfe00-f54c-446d-bc90-d18f24f51f82\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" Nov 24 17:12:32 crc kubenswrapper[4777]: I1124 17:12:32.186587 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ffbfe00-f54c-446d-bc90-d18f24f51f82-util\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c\" (UID: \"1ffbfe00-f54c-446d-bc90-d18f24f51f82\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" Nov 24 17:12:32 crc kubenswrapper[4777]: I1124 17:12:32.186658 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7knx\" (UniqueName: \"kubernetes.io/projected/1ffbfe00-f54c-446d-bc90-d18f24f51f82-kube-api-access-n7knx\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c\" (UID: \"1ffbfe00-f54c-446d-bc90-d18f24f51f82\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" Nov 24 17:12:32 crc kubenswrapper[4777]: I1124 17:12:32.186688 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ffbfe00-f54c-446d-bc90-d18f24f51f82-bundle\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c\" (UID: \"1ffbfe00-f54c-446d-bc90-d18f24f51f82\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" Nov 24 17:12:32 crc kubenswrapper[4777]: I1124 17:12:32.187204 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ffbfe00-f54c-446d-bc90-d18f24f51f82-util\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c\" (UID: \"1ffbfe00-f54c-446d-bc90-d18f24f51f82\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" Nov 24 17:12:32 crc kubenswrapper[4777]: I1124 17:12:32.187255 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ffbfe00-f54c-446d-bc90-d18f24f51f82-bundle\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c\" (UID: \"1ffbfe00-f54c-446d-bc90-d18f24f51f82\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" Nov 24 17:12:32 crc kubenswrapper[4777]: I1124 17:12:32.209904 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7knx\" (UniqueName: \"kubernetes.io/projected/1ffbfe00-f54c-446d-bc90-d18f24f51f82-kube-api-access-n7knx\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c\" (UID: \"1ffbfe00-f54c-446d-bc90-d18f24f51f82\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" Nov 24 17:12:32 crc kubenswrapper[4777]: I1124 17:12:32.304201 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" Nov 24 17:12:32 crc kubenswrapper[4777]: I1124 17:12:32.797241 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c"] Nov 24 17:12:32 crc kubenswrapper[4777]: I1124 17:12:32.895552 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" event={"ID":"1ffbfe00-f54c-446d-bc90-d18f24f51f82","Type":"ContainerStarted","Data":"e20ea5b792aeb2ad5a4e7cb82a5772a72cb45f6f0e6c5fd7bcad2b3b28b03779"} Nov 24 17:12:33 crc kubenswrapper[4777]: I1124 17:12:33.910888 4777 generic.go:334] "Generic (PLEG): container finished" podID="1ffbfe00-f54c-446d-bc90-d18f24f51f82" containerID="fcc4653e139355673a04b94069f66c131b38cdd8a0481fe97ea19c72f9815098" exitCode=0 Nov 24 17:12:33 crc kubenswrapper[4777]: I1124 17:12:33.910947 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" event={"ID":"1ffbfe00-f54c-446d-bc90-d18f24f51f82","Type":"ContainerDied","Data":"fcc4653e139355673a04b94069f66c131b38cdd8a0481fe97ea19c72f9815098"} Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.466634 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx"] Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.467677 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.470595 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.470629 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.471209 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.471918 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-vkfqz" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.472031 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.472672 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.486976 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx"] Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.516652 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d1c8f636-6ede-4469-b270-63d3823d6507-webhook-cert\") pod \"loki-operator-controller-manager-6ffccfff7d-t8ltx\" (UID: \"d1c8f636-6ede-4469-b270-63d3823d6507\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.516698 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/d1c8f636-6ede-4469-b270-63d3823d6507-manager-config\") pod \"loki-operator-controller-manager-6ffccfff7d-t8ltx\" (UID: \"d1c8f636-6ede-4469-b270-63d3823d6507\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.516751 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1c8f636-6ede-4469-b270-63d3823d6507-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-6ffccfff7d-t8ltx\" (UID: \"d1c8f636-6ede-4469-b270-63d3823d6507\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.516927 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfxzh\" (UniqueName: \"kubernetes.io/projected/d1c8f636-6ede-4469-b270-63d3823d6507-kube-api-access-sfxzh\") pod \"loki-operator-controller-manager-6ffccfff7d-t8ltx\" (UID: \"d1c8f636-6ede-4469-b270-63d3823d6507\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.517007 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d1c8f636-6ede-4469-b270-63d3823d6507-apiservice-cert\") pod \"loki-operator-controller-manager-6ffccfff7d-t8ltx\" (UID: \"d1c8f636-6ede-4469-b270-63d3823d6507\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.617768 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d1c8f636-6ede-4469-b270-63d3823d6507-apiservice-cert\") pod \"loki-operator-controller-manager-6ffccfff7d-t8ltx\" (UID: \"d1c8f636-6ede-4469-b270-63d3823d6507\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.617831 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d1c8f636-6ede-4469-b270-63d3823d6507-webhook-cert\") pod \"loki-operator-controller-manager-6ffccfff7d-t8ltx\" (UID: \"d1c8f636-6ede-4469-b270-63d3823d6507\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.617857 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/d1c8f636-6ede-4469-b270-63d3823d6507-manager-config\") pod \"loki-operator-controller-manager-6ffccfff7d-t8ltx\" (UID: \"d1c8f636-6ede-4469-b270-63d3823d6507\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.617891 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1c8f636-6ede-4469-b270-63d3823d6507-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-6ffccfff7d-t8ltx\" (UID: \"d1c8f636-6ede-4469-b270-63d3823d6507\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.617963 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfxzh\" (UniqueName: \"kubernetes.io/projected/d1c8f636-6ede-4469-b270-63d3823d6507-kube-api-access-sfxzh\") pod \"loki-operator-controller-manager-6ffccfff7d-t8ltx\" (UID: \"d1c8f636-6ede-4469-b270-63d3823d6507\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.619131 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/d1c8f636-6ede-4469-b270-63d3823d6507-manager-config\") pod \"loki-operator-controller-manager-6ffccfff7d-t8ltx\" (UID: \"d1c8f636-6ede-4469-b270-63d3823d6507\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.626493 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d1c8f636-6ede-4469-b270-63d3823d6507-webhook-cert\") pod \"loki-operator-controller-manager-6ffccfff7d-t8ltx\" (UID: \"d1c8f636-6ede-4469-b270-63d3823d6507\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.627693 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1c8f636-6ede-4469-b270-63d3823d6507-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-6ffccfff7d-t8ltx\" (UID: \"d1c8f636-6ede-4469-b270-63d3823d6507\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.643617 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfxzh\" (UniqueName: \"kubernetes.io/projected/d1c8f636-6ede-4469-b270-63d3823d6507-kube-api-access-sfxzh\") pod \"loki-operator-controller-manager-6ffccfff7d-t8ltx\" (UID: \"d1c8f636-6ede-4469-b270-63d3823d6507\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.644571 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d1c8f636-6ede-4469-b270-63d3823d6507-apiservice-cert\") pod \"loki-operator-controller-manager-6ffccfff7d-t8ltx\" (UID: \"d1c8f636-6ede-4469-b270-63d3823d6507\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.781617 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:34 crc kubenswrapper[4777]: I1124 17:12:34.977272 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx"] Nov 24 17:12:35 crc kubenswrapper[4777]: I1124 17:12:35.929583 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" event={"ID":"d1c8f636-6ede-4469-b270-63d3823d6507","Type":"ContainerStarted","Data":"07248405f145221cfe79e03b61b5e9009a11bba8a13e5e1e8b6af68106cbf42a"} Nov 24 17:12:36 crc kubenswrapper[4777]: I1124 17:12:36.939683 4777 generic.go:334] "Generic (PLEG): container finished" podID="1ffbfe00-f54c-446d-bc90-d18f24f51f82" containerID="121eab6e8c5fa89c8fb52466633f1b3508b6727a817e3304f8a9048ff6dcef05" exitCode=0 Nov 24 17:12:36 crc kubenswrapper[4777]: I1124 17:12:36.940878 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" event={"ID":"1ffbfe00-f54c-446d-bc90-d18f24f51f82","Type":"ContainerDied","Data":"121eab6e8c5fa89c8fb52466633f1b3508b6727a817e3304f8a9048ff6dcef05"} Nov 24 17:12:37 crc kubenswrapper[4777]: I1124 17:12:37.948524 4777 generic.go:334] "Generic (PLEG): container finished" podID="1ffbfe00-f54c-446d-bc90-d18f24f51f82" containerID="474abfdc7a9a5e3f40194e7de451feffa6c234d0e665ded50ec7a391a8ad7614" exitCode=0 Nov 24 17:12:37 crc kubenswrapper[4777]: I1124 17:12:37.948811 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" event={"ID":"1ffbfe00-f54c-446d-bc90-d18f24f51f82","Type":"ContainerDied","Data":"474abfdc7a9a5e3f40194e7de451feffa6c234d0e665ded50ec7a391a8ad7614"} Nov 24 17:12:39 crc kubenswrapper[4777]: I1124 17:12:39.325650 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" Nov 24 17:12:39 crc kubenswrapper[4777]: I1124 17:12:39.498884 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ffbfe00-f54c-446d-bc90-d18f24f51f82-util\") pod \"1ffbfe00-f54c-446d-bc90-d18f24f51f82\" (UID: \"1ffbfe00-f54c-446d-bc90-d18f24f51f82\") " Nov 24 17:12:39 crc kubenswrapper[4777]: I1124 17:12:39.499046 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7knx\" (UniqueName: \"kubernetes.io/projected/1ffbfe00-f54c-446d-bc90-d18f24f51f82-kube-api-access-n7knx\") pod \"1ffbfe00-f54c-446d-bc90-d18f24f51f82\" (UID: \"1ffbfe00-f54c-446d-bc90-d18f24f51f82\") " Nov 24 17:12:39 crc kubenswrapper[4777]: I1124 17:12:39.499121 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ffbfe00-f54c-446d-bc90-d18f24f51f82-bundle\") pod \"1ffbfe00-f54c-446d-bc90-d18f24f51f82\" (UID: \"1ffbfe00-f54c-446d-bc90-d18f24f51f82\") " Nov 24 17:12:39 crc kubenswrapper[4777]: I1124 17:12:39.501520 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ffbfe00-f54c-446d-bc90-d18f24f51f82-bundle" (OuterVolumeSpecName: "bundle") pod "1ffbfe00-f54c-446d-bc90-d18f24f51f82" (UID: "1ffbfe00-f54c-446d-bc90-d18f24f51f82"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:12:39 crc kubenswrapper[4777]: I1124 17:12:39.506337 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ffbfe00-f54c-446d-bc90-d18f24f51f82-kube-api-access-n7knx" (OuterVolumeSpecName: "kube-api-access-n7knx") pod "1ffbfe00-f54c-446d-bc90-d18f24f51f82" (UID: "1ffbfe00-f54c-446d-bc90-d18f24f51f82"). InnerVolumeSpecName "kube-api-access-n7knx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:12:39 crc kubenswrapper[4777]: I1124 17:12:39.511181 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ffbfe00-f54c-446d-bc90-d18f24f51f82-util" (OuterVolumeSpecName: "util") pod "1ffbfe00-f54c-446d-bc90-d18f24f51f82" (UID: "1ffbfe00-f54c-446d-bc90-d18f24f51f82"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:12:39 crc kubenswrapper[4777]: I1124 17:12:39.600909 4777 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ffbfe00-f54c-446d-bc90-d18f24f51f82-util\") on node \"crc\" DevicePath \"\"" Nov 24 17:12:39 crc kubenswrapper[4777]: I1124 17:12:39.600963 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7knx\" (UniqueName: \"kubernetes.io/projected/1ffbfe00-f54c-446d-bc90-d18f24f51f82-kube-api-access-n7knx\") on node \"crc\" DevicePath \"\"" Nov 24 17:12:39 crc kubenswrapper[4777]: I1124 17:12:39.600996 4777 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ffbfe00-f54c-446d-bc90-d18f24f51f82-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:12:39 crc kubenswrapper[4777]: I1124 17:12:39.970081 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" event={"ID":"1ffbfe00-f54c-446d-bc90-d18f24f51f82","Type":"ContainerDied","Data":"e20ea5b792aeb2ad5a4e7cb82a5772a72cb45f6f0e6c5fd7bcad2b3b28b03779"} Nov 24 17:12:39 crc kubenswrapper[4777]: I1124 17:12:39.970119 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e20ea5b792aeb2ad5a4e7cb82a5772a72cb45f6f0e6c5fd7bcad2b3b28b03779" Nov 24 17:12:39 crc kubenswrapper[4777]: I1124 17:12:39.970165 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c" Nov 24 17:12:40 crc kubenswrapper[4777]: I1124 17:12:40.979728 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" event={"ID":"d1c8f636-6ede-4469-b270-63d3823d6507","Type":"ContainerStarted","Data":"82566ca62e20948dbfed099707d3aaa9d858f47cc5fe620065ad2fd724a56c38"} Nov 24 17:12:44 crc kubenswrapper[4777]: I1124 17:12:44.270509 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:12:44 crc kubenswrapper[4777]: I1124 17:12:44.270571 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:12:48 crc kubenswrapper[4777]: I1124 17:12:48.034865 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" event={"ID":"d1c8f636-6ede-4469-b270-63d3823d6507","Type":"ContainerStarted","Data":"73f52a1d4391134c85eef1e9c1755c2fe28a70d1c7a9db35ab5792b71a773b5c"} Nov 24 17:12:48 crc kubenswrapper[4777]: I1124 17:12:48.035263 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:48 crc kubenswrapper[4777]: I1124 17:12:48.039900 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" Nov 24 17:12:48 crc kubenswrapper[4777]: I1124 17:12:48.067933 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-6ffccfff7d-t8ltx" podStartSLOduration=1.435596302 podStartE2EDuration="14.067902952s" podCreationTimestamp="2025-11-24 17:12:34 +0000 UTC" firstStartedPulling="2025-11-24 17:12:34.985665998 +0000 UTC m=+713.144861057" lastFinishedPulling="2025-11-24 17:12:47.617972638 +0000 UTC m=+725.777167707" observedRunningTime="2025-11-24 17:12:48.065640174 +0000 UTC m=+726.224835253" watchObservedRunningTime="2025-11-24 17:12:48.067902952 +0000 UTC m=+726.227098041" Nov 24 17:13:14 crc kubenswrapper[4777]: I1124 17:13:14.270795 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:13:14 crc kubenswrapper[4777]: I1124 17:13:14.271522 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:13:14 crc kubenswrapper[4777]: I1124 17:13:14.271610 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:13:14 crc kubenswrapper[4777]: I1124 17:13:14.272309 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a0f6b6c1a5fc2ed6243c1d079f9f074d4a99e988147412f7b07f358cd4f1c26d"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:13:14 crc kubenswrapper[4777]: I1124 17:13:14.272406 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://a0f6b6c1a5fc2ed6243c1d079f9f074d4a99e988147412f7b07f358cd4f1c26d" gracePeriod=600 Nov 24 17:13:15 crc kubenswrapper[4777]: I1124 17:13:15.229448 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="a0f6b6c1a5fc2ed6243c1d079f9f074d4a99e988147412f7b07f358cd4f1c26d" exitCode=0 Nov 24 17:13:15 crc kubenswrapper[4777]: I1124 17:13:15.229529 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"a0f6b6c1a5fc2ed6243c1d079f9f074d4a99e988147412f7b07f358cd4f1c26d"} Nov 24 17:13:15 crc kubenswrapper[4777]: I1124 17:13:15.230152 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"90d22c4d0e8b2ba20584ba22b6a107d623c9ded1ab193fb10e43b251d31391cc"} Nov 24 17:13:15 crc kubenswrapper[4777]: I1124 17:13:15.230183 4777 scope.go:117] "RemoveContainer" containerID="eb08dd4022cfe7bafee35dcfc7cb94f0a8332bb8d48affe8af298f1e887ff112" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.409678 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww"] Nov 24 17:13:20 crc kubenswrapper[4777]: E1124 17:13:20.410349 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ffbfe00-f54c-446d-bc90-d18f24f51f82" containerName="extract" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.410363 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ffbfe00-f54c-446d-bc90-d18f24f51f82" containerName="extract" Nov 24 17:13:20 crc kubenswrapper[4777]: E1124 17:13:20.410375 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ffbfe00-f54c-446d-bc90-d18f24f51f82" containerName="pull" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.410383 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ffbfe00-f54c-446d-bc90-d18f24f51f82" containerName="pull" Nov 24 17:13:20 crc kubenswrapper[4777]: E1124 17:13:20.410398 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ffbfe00-f54c-446d-bc90-d18f24f51f82" containerName="util" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.410406 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ffbfe00-f54c-446d-bc90-d18f24f51f82" containerName="util" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.410530 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ffbfe00-f54c-446d-bc90-d18f24f51f82" containerName="extract" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.411459 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.413954 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.420514 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vs5l\" (UniqueName: \"kubernetes.io/projected/213b5cf0-7684-4331-b5be-28fd188e0db3-kube-api-access-7vs5l\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww\" (UID: \"213b5cf0-7684-4331-b5be-28fd188e0db3\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.420636 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/213b5cf0-7684-4331-b5be-28fd188e0db3-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww\" (UID: \"213b5cf0-7684-4331-b5be-28fd188e0db3\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.420846 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/213b5cf0-7684-4331-b5be-28fd188e0db3-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww\" (UID: \"213b5cf0-7684-4331-b5be-28fd188e0db3\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.422759 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww"] Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.521603 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/213b5cf0-7684-4331-b5be-28fd188e0db3-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww\" (UID: \"213b5cf0-7684-4331-b5be-28fd188e0db3\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.521700 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/213b5cf0-7684-4331-b5be-28fd188e0db3-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww\" (UID: \"213b5cf0-7684-4331-b5be-28fd188e0db3\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.521740 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vs5l\" (UniqueName: \"kubernetes.io/projected/213b5cf0-7684-4331-b5be-28fd188e0db3-kube-api-access-7vs5l\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww\" (UID: \"213b5cf0-7684-4331-b5be-28fd188e0db3\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.522285 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/213b5cf0-7684-4331-b5be-28fd188e0db3-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww\" (UID: \"213b5cf0-7684-4331-b5be-28fd188e0db3\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.522323 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/213b5cf0-7684-4331-b5be-28fd188e0db3-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww\" (UID: \"213b5cf0-7684-4331-b5be-28fd188e0db3\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.559620 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vs5l\" (UniqueName: \"kubernetes.io/projected/213b5cf0-7684-4331-b5be-28fd188e0db3-kube-api-access-7vs5l\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww\" (UID: \"213b5cf0-7684-4331-b5be-28fd188e0db3\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" Nov 24 17:13:20 crc kubenswrapper[4777]: I1124 17:13:20.762011 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" Nov 24 17:13:21 crc kubenswrapper[4777]: I1124 17:13:21.029969 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww"] Nov 24 17:13:21 crc kubenswrapper[4777]: I1124 17:13:21.314424 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" event={"ID":"213b5cf0-7684-4331-b5be-28fd188e0db3","Type":"ContainerStarted","Data":"4c4ae51ead7282f24c432e3c770babc98415fb19a723a3720b1c840308629fef"} Nov 24 17:13:21 crc kubenswrapper[4777]: I1124 17:13:21.314486 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" event={"ID":"213b5cf0-7684-4331-b5be-28fd188e0db3","Type":"ContainerStarted","Data":"472dd9a4508002c4cf978ed37a4105ebc71e80ffcef8d8843ef7473569f01be4"} Nov 24 17:13:22 crc kubenswrapper[4777]: I1124 17:13:22.345071 4777 generic.go:334] "Generic (PLEG): container finished" podID="213b5cf0-7684-4331-b5be-28fd188e0db3" containerID="4c4ae51ead7282f24c432e3c770babc98415fb19a723a3720b1c840308629fef" exitCode=0 Nov 24 17:13:22 crc kubenswrapper[4777]: I1124 17:13:22.345136 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" event={"ID":"213b5cf0-7684-4331-b5be-28fd188e0db3","Type":"ContainerDied","Data":"4c4ae51ead7282f24c432e3c770babc98415fb19a723a3720b1c840308629fef"} Nov 24 17:13:23 crc kubenswrapper[4777]: I1124 17:13:23.354532 4777 generic.go:334] "Generic (PLEG): container finished" podID="213b5cf0-7684-4331-b5be-28fd188e0db3" containerID="d65dc12d9417f74e81096144e2d605bea369cc442a132f4c1bcd56d4312f7a8a" exitCode=0 Nov 24 17:13:23 crc kubenswrapper[4777]: I1124 17:13:23.354804 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" event={"ID":"213b5cf0-7684-4331-b5be-28fd188e0db3","Type":"ContainerDied","Data":"d65dc12d9417f74e81096144e2d605bea369cc442a132f4c1bcd56d4312f7a8a"} Nov 24 17:13:24 crc kubenswrapper[4777]: I1124 17:13:24.364844 4777 generic.go:334] "Generic (PLEG): container finished" podID="213b5cf0-7684-4331-b5be-28fd188e0db3" containerID="912a8796a4b6c7f734039909634da4ea4545f0236034f8368c8814a3fe47f50c" exitCode=0 Nov 24 17:13:24 crc kubenswrapper[4777]: I1124 17:13:24.364911 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" event={"ID":"213b5cf0-7684-4331-b5be-28fd188e0db3","Type":"ContainerDied","Data":"912a8796a4b6c7f734039909634da4ea4545f0236034f8368c8814a3fe47f50c"} Nov 24 17:13:25 crc kubenswrapper[4777]: I1124 17:13:25.689019 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" Nov 24 17:13:25 crc kubenswrapper[4777]: I1124 17:13:25.738505 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/213b5cf0-7684-4331-b5be-28fd188e0db3-bundle\") pod \"213b5cf0-7684-4331-b5be-28fd188e0db3\" (UID: \"213b5cf0-7684-4331-b5be-28fd188e0db3\") " Nov 24 17:13:25 crc kubenswrapper[4777]: I1124 17:13:25.738579 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vs5l\" (UniqueName: \"kubernetes.io/projected/213b5cf0-7684-4331-b5be-28fd188e0db3-kube-api-access-7vs5l\") pod \"213b5cf0-7684-4331-b5be-28fd188e0db3\" (UID: \"213b5cf0-7684-4331-b5be-28fd188e0db3\") " Nov 24 17:13:25 crc kubenswrapper[4777]: I1124 17:13:25.738663 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/213b5cf0-7684-4331-b5be-28fd188e0db3-util\") pod \"213b5cf0-7684-4331-b5be-28fd188e0db3\" (UID: \"213b5cf0-7684-4331-b5be-28fd188e0db3\") " Nov 24 17:13:25 crc kubenswrapper[4777]: I1124 17:13:25.740261 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/213b5cf0-7684-4331-b5be-28fd188e0db3-bundle" (OuterVolumeSpecName: "bundle") pod "213b5cf0-7684-4331-b5be-28fd188e0db3" (UID: "213b5cf0-7684-4331-b5be-28fd188e0db3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:13:25 crc kubenswrapper[4777]: I1124 17:13:25.749430 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/213b5cf0-7684-4331-b5be-28fd188e0db3-kube-api-access-7vs5l" (OuterVolumeSpecName: "kube-api-access-7vs5l") pod "213b5cf0-7684-4331-b5be-28fd188e0db3" (UID: "213b5cf0-7684-4331-b5be-28fd188e0db3"). InnerVolumeSpecName "kube-api-access-7vs5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:13:25 crc kubenswrapper[4777]: I1124 17:13:25.763508 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/213b5cf0-7684-4331-b5be-28fd188e0db3-util" (OuterVolumeSpecName: "util") pod "213b5cf0-7684-4331-b5be-28fd188e0db3" (UID: "213b5cf0-7684-4331-b5be-28fd188e0db3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:13:25 crc kubenswrapper[4777]: I1124 17:13:25.840132 4777 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/213b5cf0-7684-4331-b5be-28fd188e0db3-util\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:25 crc kubenswrapper[4777]: I1124 17:13:25.840179 4777 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/213b5cf0-7684-4331-b5be-28fd188e0db3-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:25 crc kubenswrapper[4777]: I1124 17:13:25.840198 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vs5l\" (UniqueName: \"kubernetes.io/projected/213b5cf0-7684-4331-b5be-28fd188e0db3-kube-api-access-7vs5l\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:26 crc kubenswrapper[4777]: I1124 17:13:26.386388 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" event={"ID":"213b5cf0-7684-4331-b5be-28fd188e0db3","Type":"ContainerDied","Data":"472dd9a4508002c4cf978ed37a4105ebc71e80ffcef8d8843ef7473569f01be4"} Nov 24 17:13:26 crc kubenswrapper[4777]: I1124 17:13:26.386815 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="472dd9a4508002c4cf978ed37a4105ebc71e80ffcef8d8843ef7473569f01be4" Nov 24 17:13:26 crc kubenswrapper[4777]: I1124 17:13:26.386504 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww" Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.494874 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-8wtx5"] Nov 24 17:13:32 crc kubenswrapper[4777]: E1124 17:13:32.495855 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="213b5cf0-7684-4331-b5be-28fd188e0db3" containerName="pull" Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.495876 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="213b5cf0-7684-4331-b5be-28fd188e0db3" containerName="pull" Nov 24 17:13:32 crc kubenswrapper[4777]: E1124 17:13:32.495900 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="213b5cf0-7684-4331-b5be-28fd188e0db3" containerName="extract" Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.495911 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="213b5cf0-7684-4331-b5be-28fd188e0db3" containerName="extract" Nov 24 17:13:32 crc kubenswrapper[4777]: E1124 17:13:32.495942 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="213b5cf0-7684-4331-b5be-28fd188e0db3" containerName="util" Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.495955 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="213b5cf0-7684-4331-b5be-28fd188e0db3" containerName="util" Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.496171 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="213b5cf0-7684-4331-b5be-28fd188e0db3" containerName="extract" Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.496784 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-8wtx5" Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.498720 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.498847 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.499279 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-klbjb" Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.523068 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-8wtx5"] Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.628501 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69fxw\" (UniqueName: \"kubernetes.io/projected/3ffb9c2f-e94c-4fda-8c82-a34c045f9435-kube-api-access-69fxw\") pod \"nmstate-operator-557fdffb88-8wtx5\" (UID: \"3ffb9c2f-e94c-4fda-8c82-a34c045f9435\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-8wtx5" Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.682700 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kd6cl"] Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.682900 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" podUID="9773b6b0-ffe7-40ed-b714-c7266b46e298" containerName="controller-manager" containerID="cri-o://da2f196fb92c46abfe2c2ef899badcea0b32cdb68db4cb7964e93ef3f06d6dca" gracePeriod=30 Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.729596 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69fxw\" (UniqueName: \"kubernetes.io/projected/3ffb9c2f-e94c-4fda-8c82-a34c045f9435-kube-api-access-69fxw\") pod \"nmstate-operator-557fdffb88-8wtx5\" (UID: \"3ffb9c2f-e94c-4fda-8c82-a34c045f9435\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-8wtx5" Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.751874 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69fxw\" (UniqueName: \"kubernetes.io/projected/3ffb9c2f-e94c-4fda-8c82-a34c045f9435-kube-api-access-69fxw\") pod \"nmstate-operator-557fdffb88-8wtx5\" (UID: \"3ffb9c2f-e94c-4fda-8c82-a34c045f9435\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-8wtx5" Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.786288 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t"] Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.786482 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" podUID="2148eaac-ce88-4048-9ae9-9ff6df2d4747" containerName="route-controller-manager" containerID="cri-o://f3f60291ec83f38106556bbf549de51d0ce0ff286edbec947edbc1a5d4c3fa66" gracePeriod=30 Nov 24 17:13:32 crc kubenswrapper[4777]: I1124 17:13:32.813463 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-8wtx5" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.076085 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-8wtx5"] Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.148123 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.253419 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-config\") pod \"9773b6b0-ffe7-40ed-b714-c7266b46e298\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.253539 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-client-ca\") pod \"9773b6b0-ffe7-40ed-b714-c7266b46e298\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.253585 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77rt5\" (UniqueName: \"kubernetes.io/projected/9773b6b0-ffe7-40ed-b714-c7266b46e298-kube-api-access-77rt5\") pod \"9773b6b0-ffe7-40ed-b714-c7266b46e298\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.253603 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-proxy-ca-bundles\") pod \"9773b6b0-ffe7-40ed-b714-c7266b46e298\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.253635 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9773b6b0-ffe7-40ed-b714-c7266b46e298-serving-cert\") pod \"9773b6b0-ffe7-40ed-b714-c7266b46e298\" (UID: \"9773b6b0-ffe7-40ed-b714-c7266b46e298\") " Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.254450 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-client-ca" (OuterVolumeSpecName: "client-ca") pod "9773b6b0-ffe7-40ed-b714-c7266b46e298" (UID: "9773b6b0-ffe7-40ed-b714-c7266b46e298"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.254548 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-config" (OuterVolumeSpecName: "config") pod "9773b6b0-ffe7-40ed-b714-c7266b46e298" (UID: "9773b6b0-ffe7-40ed-b714-c7266b46e298"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.254594 4777 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.255095 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9773b6b0-ffe7-40ed-b714-c7266b46e298" (UID: "9773b6b0-ffe7-40ed-b714-c7266b46e298"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.259694 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9773b6b0-ffe7-40ed-b714-c7266b46e298-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9773b6b0-ffe7-40ed-b714-c7266b46e298" (UID: "9773b6b0-ffe7-40ed-b714-c7266b46e298"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.281245 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9773b6b0-ffe7-40ed-b714-c7266b46e298-kube-api-access-77rt5" (OuterVolumeSpecName: "kube-api-access-77rt5") pod "9773b6b0-ffe7-40ed-b714-c7266b46e298" (UID: "9773b6b0-ffe7-40ed-b714-c7266b46e298"). InnerVolumeSpecName "kube-api-access-77rt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.328498 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.355173 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2148eaac-ce88-4048-9ae9-9ff6df2d4747-client-ca\") pod \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.355230 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2148eaac-ce88-4048-9ae9-9ff6df2d4747-serving-cert\") pod \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.355260 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2148eaac-ce88-4048-9ae9-9ff6df2d4747-config\") pod \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.355321 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6r9w\" (UniqueName: \"kubernetes.io/projected/2148eaac-ce88-4048-9ae9-9ff6df2d4747-kube-api-access-j6r9w\") pod \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\" (UID: \"2148eaac-ce88-4048-9ae9-9ff6df2d4747\") " Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.355603 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77rt5\" (UniqueName: \"kubernetes.io/projected/9773b6b0-ffe7-40ed-b714-c7266b46e298-kube-api-access-77rt5\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.355619 4777 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.355628 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9773b6b0-ffe7-40ed-b714-c7266b46e298-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.355637 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9773b6b0-ffe7-40ed-b714-c7266b46e298-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.355809 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2148eaac-ce88-4048-9ae9-9ff6df2d4747-client-ca" (OuterVolumeSpecName: "client-ca") pod "2148eaac-ce88-4048-9ae9-9ff6df2d4747" (UID: "2148eaac-ce88-4048-9ae9-9ff6df2d4747"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.355919 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2148eaac-ce88-4048-9ae9-9ff6df2d4747-config" (OuterVolumeSpecName: "config") pod "2148eaac-ce88-4048-9ae9-9ff6df2d4747" (UID: "2148eaac-ce88-4048-9ae9-9ff6df2d4747"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.358475 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2148eaac-ce88-4048-9ae9-9ff6df2d4747-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2148eaac-ce88-4048-9ae9-9ff6df2d4747" (UID: "2148eaac-ce88-4048-9ae9-9ff6df2d4747"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.361391 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2148eaac-ce88-4048-9ae9-9ff6df2d4747-kube-api-access-j6r9w" (OuterVolumeSpecName: "kube-api-access-j6r9w") pod "2148eaac-ce88-4048-9ae9-9ff6df2d4747" (UID: "2148eaac-ce88-4048-9ae9-9ff6df2d4747"). InnerVolumeSpecName "kube-api-access-j6r9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.437275 4777 generic.go:334] "Generic (PLEG): container finished" podID="9773b6b0-ffe7-40ed-b714-c7266b46e298" containerID="da2f196fb92c46abfe2c2ef899badcea0b32cdb68db4cb7964e93ef3f06d6dca" exitCode=0 Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.437335 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.437352 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" event={"ID":"9773b6b0-ffe7-40ed-b714-c7266b46e298","Type":"ContainerDied","Data":"da2f196fb92c46abfe2c2ef899badcea0b32cdb68db4cb7964e93ef3f06d6dca"} Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.437384 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kd6cl" event={"ID":"9773b6b0-ffe7-40ed-b714-c7266b46e298","Type":"ContainerDied","Data":"b4d91654bd09070f705540d28bab0d20b5c67cdfc36852d890fe776f6a4340d4"} Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.437407 4777 scope.go:117] "RemoveContainer" containerID="da2f196fb92c46abfe2c2ef899badcea0b32cdb68db4cb7964e93ef3f06d6dca" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.440520 4777 generic.go:334] "Generic (PLEG): container finished" podID="2148eaac-ce88-4048-9ae9-9ff6df2d4747" containerID="f3f60291ec83f38106556bbf549de51d0ce0ff286edbec947edbc1a5d4c3fa66" exitCode=0 Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.440601 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.440607 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" event={"ID":"2148eaac-ce88-4048-9ae9-9ff6df2d4747","Type":"ContainerDied","Data":"f3f60291ec83f38106556bbf549de51d0ce0ff286edbec947edbc1a5d4c3fa66"} Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.440721 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t" event={"ID":"2148eaac-ce88-4048-9ae9-9ff6df2d4747","Type":"ContainerDied","Data":"fe754e761e4badd714b65f963f8488c330a3b82534ca3aab6610b311eb7192b5"} Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.441713 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-8wtx5" event={"ID":"3ffb9c2f-e94c-4fda-8c82-a34c045f9435","Type":"ContainerStarted","Data":"0f2098325794b06740935a696d001aeec26d1f98849c0fc75f1820d1887bd371"} Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.454515 4777 scope.go:117] "RemoveContainer" containerID="da2f196fb92c46abfe2c2ef899badcea0b32cdb68db4cb7964e93ef3f06d6dca" Nov 24 17:13:33 crc kubenswrapper[4777]: E1124 17:13:33.454880 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da2f196fb92c46abfe2c2ef899badcea0b32cdb68db4cb7964e93ef3f06d6dca\": container with ID starting with da2f196fb92c46abfe2c2ef899badcea0b32cdb68db4cb7964e93ef3f06d6dca not found: ID does not exist" containerID="da2f196fb92c46abfe2c2ef899badcea0b32cdb68db4cb7964e93ef3f06d6dca" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.454924 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da2f196fb92c46abfe2c2ef899badcea0b32cdb68db4cb7964e93ef3f06d6dca"} err="failed to get container status \"da2f196fb92c46abfe2c2ef899badcea0b32cdb68db4cb7964e93ef3f06d6dca\": rpc error: code = NotFound desc = could not find container \"da2f196fb92c46abfe2c2ef899badcea0b32cdb68db4cb7964e93ef3f06d6dca\": container with ID starting with da2f196fb92c46abfe2c2ef899badcea0b32cdb68db4cb7964e93ef3f06d6dca not found: ID does not exist" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.454952 4777 scope.go:117] "RemoveContainer" containerID="f3f60291ec83f38106556bbf549de51d0ce0ff286edbec947edbc1a5d4c3fa66" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.456292 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6r9w\" (UniqueName: \"kubernetes.io/projected/2148eaac-ce88-4048-9ae9-9ff6df2d4747-kube-api-access-j6r9w\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.456309 4777 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2148eaac-ce88-4048-9ae9-9ff6df2d4747-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.456319 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2148eaac-ce88-4048-9ae9-9ff6df2d4747-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.456327 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2148eaac-ce88-4048-9ae9-9ff6df2d4747-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.476156 4777 scope.go:117] "RemoveContainer" containerID="f3f60291ec83f38106556bbf549de51d0ce0ff286edbec947edbc1a5d4c3fa66" Nov 24 17:13:33 crc kubenswrapper[4777]: E1124 17:13:33.476714 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3f60291ec83f38106556bbf549de51d0ce0ff286edbec947edbc1a5d4c3fa66\": container with ID starting with f3f60291ec83f38106556bbf549de51d0ce0ff286edbec947edbc1a5d4c3fa66 not found: ID does not exist" containerID="f3f60291ec83f38106556bbf549de51d0ce0ff286edbec947edbc1a5d4c3fa66" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.476792 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3f60291ec83f38106556bbf549de51d0ce0ff286edbec947edbc1a5d4c3fa66"} err="failed to get container status \"f3f60291ec83f38106556bbf549de51d0ce0ff286edbec947edbc1a5d4c3fa66\": rpc error: code = NotFound desc = could not find container \"f3f60291ec83f38106556bbf549de51d0ce0ff286edbec947edbc1a5d4c3fa66\": container with ID starting with f3f60291ec83f38106556bbf549de51d0ce0ff286edbec947edbc1a5d4c3fa66 not found: ID does not exist" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.483252 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kd6cl"] Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.486269 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kd6cl"] Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.490098 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t"] Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.493346 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zzd7t"] Nov 24 17:13:33 crc kubenswrapper[4777]: E1124 17:13:33.570237 4777 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2148eaac_ce88_4048_9ae9_9ff6df2d4747.slice/crio-fe754e761e4badd714b65f963f8488c330a3b82534ca3aab6610b311eb7192b5\": RecentStats: unable to find data in memory cache]" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.994123 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d"] Nov 24 17:13:33 crc kubenswrapper[4777]: E1124 17:13:33.995044 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9773b6b0-ffe7-40ed-b714-c7266b46e298" containerName="controller-manager" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.995090 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9773b6b0-ffe7-40ed-b714-c7266b46e298" containerName="controller-manager" Nov 24 17:13:33 crc kubenswrapper[4777]: E1124 17:13:33.995122 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2148eaac-ce88-4048-9ae9-9ff6df2d4747" containerName="route-controller-manager" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.995146 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="2148eaac-ce88-4048-9ae9-9ff6df2d4747" containerName="route-controller-manager" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.995412 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9773b6b0-ffe7-40ed-b714-c7266b46e298" containerName="controller-manager" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.995465 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="2148eaac-ce88-4048-9ae9-9ff6df2d4747" containerName="route-controller-manager" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.996277 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.998009 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5f579785f-bddj2"] Nov 24 17:13:33 crc kubenswrapper[4777]: I1124 17:13:33.998820 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.000569 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.003403 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.003805 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.003874 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.004007 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.004472 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.005260 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.005340 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.006018 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.006345 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.006512 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.006606 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.019150 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.022035 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f579785f-bddj2"] Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.033209 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d"] Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.063595 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1407cd97-3ca4-4f29-a6be-700428acce49-serving-cert\") pod \"controller-manager-5f579785f-bddj2\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.063683 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-proxy-ca-bundles\") pod \"controller-manager-5f579785f-bddj2\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.067169 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqtjb\" (UniqueName: \"kubernetes.io/projected/1407cd97-3ca4-4f29-a6be-700428acce49-kube-api-access-gqtjb\") pod \"controller-manager-5f579785f-bddj2\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.067298 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-client-ca\") pod \"controller-manager-5f579785f-bddj2\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.067353 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-config\") pod \"controller-manager-5f579785f-bddj2\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.067421 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-serving-cert\") pod \"route-controller-manager-5548cdb5fc-f9v4d\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.067458 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbjcc\" (UniqueName: \"kubernetes.io/projected/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-kube-api-access-fbjcc\") pod \"route-controller-manager-5548cdb5fc-f9v4d\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.067513 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-config\") pod \"route-controller-manager-5548cdb5fc-f9v4d\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.067549 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-client-ca\") pod \"route-controller-manager-5548cdb5fc-f9v4d\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.168922 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-client-ca\") pod \"controller-manager-5f579785f-bddj2\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.168998 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-config\") pod \"controller-manager-5f579785f-bddj2\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.169032 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-serving-cert\") pod \"route-controller-manager-5548cdb5fc-f9v4d\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.169051 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbjcc\" (UniqueName: \"kubernetes.io/projected/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-kube-api-access-fbjcc\") pod \"route-controller-manager-5548cdb5fc-f9v4d\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.169084 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-config\") pod \"route-controller-manager-5548cdb5fc-f9v4d\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.169101 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-client-ca\") pod \"route-controller-manager-5548cdb5fc-f9v4d\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.169127 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1407cd97-3ca4-4f29-a6be-700428acce49-serving-cert\") pod \"controller-manager-5f579785f-bddj2\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.169152 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-proxy-ca-bundles\") pod \"controller-manager-5f579785f-bddj2\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.169173 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqtjb\" (UniqueName: \"kubernetes.io/projected/1407cd97-3ca4-4f29-a6be-700428acce49-kube-api-access-gqtjb\") pod \"controller-manager-5f579785f-bddj2\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.170220 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-client-ca\") pod \"controller-manager-5f579785f-bddj2\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.170220 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-client-ca\") pod \"route-controller-manager-5548cdb5fc-f9v4d\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.170361 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-config\") pod \"route-controller-manager-5548cdb5fc-f9v4d\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.170703 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-config\") pod \"controller-manager-5f579785f-bddj2\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.170805 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-proxy-ca-bundles\") pod \"controller-manager-5f579785f-bddj2\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.173629 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1407cd97-3ca4-4f29-a6be-700428acce49-serving-cert\") pod \"controller-manager-5f579785f-bddj2\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.185288 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-serving-cert\") pod \"route-controller-manager-5548cdb5fc-f9v4d\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.200037 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqtjb\" (UniqueName: \"kubernetes.io/projected/1407cd97-3ca4-4f29-a6be-700428acce49-kube-api-access-gqtjb\") pod \"controller-manager-5f579785f-bddj2\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.200513 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbjcc\" (UniqueName: \"kubernetes.io/projected/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-kube-api-access-fbjcc\") pod \"route-controller-manager-5548cdb5fc-f9v4d\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.266284 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5f579785f-bddj2"] Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.266664 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.289525 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d"] Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.289949 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.525685 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d"] Nov 24 17:13:34 crc kubenswrapper[4777]: W1124 17:13:34.528543 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfa08af0_0c1f_446a_9a55_4aeb0cdb7179.slice/crio-c42ce4aca5287466cd92cb6a62d0c7922c999ee4c7e1a2ce5049264e60dbfbb7 WatchSource:0}: Error finding container c42ce4aca5287466cd92cb6a62d0c7922c999ee4c7e1a2ce5049264e60dbfbb7: Status 404 returned error can't find the container with id c42ce4aca5287466cd92cb6a62d0c7922c999ee4c7e1a2ce5049264e60dbfbb7 Nov 24 17:13:34 crc kubenswrapper[4777]: I1124 17:13:34.715766 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5f579785f-bddj2"] Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.251490 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2148eaac-ce88-4048-9ae9-9ff6df2d4747" path="/var/lib/kubelet/pods/2148eaac-ce88-4048-9ae9-9ff6df2d4747/volumes" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.252148 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9773b6b0-ffe7-40ed-b714-c7266b46e298" path="/var/lib/kubelet/pods/9773b6b0-ffe7-40ed-b714-c7266b46e298/volumes" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.484731 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" event={"ID":"1407cd97-3ca4-4f29-a6be-700428acce49","Type":"ContainerStarted","Data":"e6e5c1264bc703abcb3f3ee75b51fcdd59b7ce26df13e617ec5ca9d372043b52"} Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.484772 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" event={"ID":"1407cd97-3ca4-4f29-a6be-700428acce49","Type":"ContainerStarted","Data":"dae93807c0e49a974fb2149a3e0bb89a10382cfef96c8d6ae853fdc368e06ad5"} Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.484863 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" podUID="1407cd97-3ca4-4f29-a6be-700428acce49" containerName="controller-manager" containerID="cri-o://e6e5c1264bc703abcb3f3ee75b51fcdd59b7ce26df13e617ec5ca9d372043b52" gracePeriod=30 Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.485314 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.490444 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.492545 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" event={"ID":"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179","Type":"ContainerStarted","Data":"edaccbc37f80bda8ea2f651d7a25c1e889c3136277f3b9d4c9515022c2de66b9"} Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.492592 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" event={"ID":"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179","Type":"ContainerStarted","Data":"c42ce4aca5287466cd92cb6a62d0c7922c999ee4c7e1a2ce5049264e60dbfbb7"} Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.492639 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" podUID="bfa08af0-0c1f-446a-9a55-4aeb0cdb7179" containerName="route-controller-manager" containerID="cri-o://edaccbc37f80bda8ea2f651d7a25c1e889c3136277f3b9d4c9515022c2de66b9" gracePeriod=30 Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.492833 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.498217 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.505201 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" podStartSLOduration=3.505184091 podStartE2EDuration="3.505184091s" podCreationTimestamp="2025-11-24 17:13:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:13:35.503327877 +0000 UTC m=+773.662522936" watchObservedRunningTime="2025-11-24 17:13:35.505184091 +0000 UTC m=+773.664379140" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.524007 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" podStartSLOduration=3.523990675 podStartE2EDuration="3.523990675s" podCreationTimestamp="2025-11-24 17:13:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:13:35.522097591 +0000 UTC m=+773.681292650" watchObservedRunningTime="2025-11-24 17:13:35.523990675 +0000 UTC m=+773.683185734" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.846742 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.870289 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv"] Nov 24 17:13:35 crc kubenswrapper[4777]: E1124 17:13:35.870552 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfa08af0-0c1f-446a-9a55-4aeb0cdb7179" containerName="route-controller-manager" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.870573 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfa08af0-0c1f-446a-9a55-4aeb0cdb7179" containerName="route-controller-manager" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.870692 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfa08af0-0c1f-446a-9a55-4aeb0cdb7179" containerName="route-controller-manager" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.871204 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.882084 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv"] Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.925245 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.994350 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-serving-cert\") pod \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.994440 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbjcc\" (UniqueName: \"kubernetes.io/projected/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-kube-api-access-fbjcc\") pod \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.994486 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-config\") pod \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.994518 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-client-ca\") pod \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\" (UID: \"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179\") " Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.994687 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpw45\" (UniqueName: \"kubernetes.io/projected/9393c0c7-756e-4be8-b863-368986fbcf0a-kube-api-access-dpw45\") pod \"route-controller-manager-695bbbd669-m87vv\" (UID: \"9393c0c7-756e-4be8-b863-368986fbcf0a\") " pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.994713 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9393c0c7-756e-4be8-b863-368986fbcf0a-serving-cert\") pod \"route-controller-manager-695bbbd669-m87vv\" (UID: \"9393c0c7-756e-4be8-b863-368986fbcf0a\") " pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.994734 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9393c0c7-756e-4be8-b863-368986fbcf0a-client-ca\") pod \"route-controller-manager-695bbbd669-m87vv\" (UID: \"9393c0c7-756e-4be8-b863-368986fbcf0a\") " pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.994751 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9393c0c7-756e-4be8-b863-368986fbcf0a-config\") pod \"route-controller-manager-695bbbd669-m87vv\" (UID: \"9393c0c7-756e-4be8-b863-368986fbcf0a\") " pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:35 crc kubenswrapper[4777]: I1124 17:13:35.996538 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-client-ca" (OuterVolumeSpecName: "client-ca") pod "bfa08af0-0c1f-446a-9a55-4aeb0cdb7179" (UID: "bfa08af0-0c1f-446a-9a55-4aeb0cdb7179"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.000036 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-config" (OuterVolumeSpecName: "config") pod "bfa08af0-0c1f-446a-9a55-4aeb0cdb7179" (UID: "bfa08af0-0c1f-446a-9a55-4aeb0cdb7179"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.004153 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bfa08af0-0c1f-446a-9a55-4aeb0cdb7179" (UID: "bfa08af0-0c1f-446a-9a55-4aeb0cdb7179"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.016492 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-kube-api-access-fbjcc" (OuterVolumeSpecName: "kube-api-access-fbjcc") pod "bfa08af0-0c1f-446a-9a55-4aeb0cdb7179" (UID: "bfa08af0-0c1f-446a-9a55-4aeb0cdb7179"). InnerVolumeSpecName "kube-api-access-fbjcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.095329 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqtjb\" (UniqueName: \"kubernetes.io/projected/1407cd97-3ca4-4f29-a6be-700428acce49-kube-api-access-gqtjb\") pod \"1407cd97-3ca4-4f29-a6be-700428acce49\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.095409 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1407cd97-3ca4-4f29-a6be-700428acce49-serving-cert\") pod \"1407cd97-3ca4-4f29-a6be-700428acce49\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.095473 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-client-ca\") pod \"1407cd97-3ca4-4f29-a6be-700428acce49\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.095516 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-proxy-ca-bundles\") pod \"1407cd97-3ca4-4f29-a6be-700428acce49\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.095557 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-config\") pod \"1407cd97-3ca4-4f29-a6be-700428acce49\" (UID: \"1407cd97-3ca4-4f29-a6be-700428acce49\") " Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.095759 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpw45\" (UniqueName: \"kubernetes.io/projected/9393c0c7-756e-4be8-b863-368986fbcf0a-kube-api-access-dpw45\") pod \"route-controller-manager-695bbbd669-m87vv\" (UID: \"9393c0c7-756e-4be8-b863-368986fbcf0a\") " pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.095786 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9393c0c7-756e-4be8-b863-368986fbcf0a-serving-cert\") pod \"route-controller-manager-695bbbd669-m87vv\" (UID: \"9393c0c7-756e-4be8-b863-368986fbcf0a\") " pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.095812 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9393c0c7-756e-4be8-b863-368986fbcf0a-client-ca\") pod \"route-controller-manager-695bbbd669-m87vv\" (UID: \"9393c0c7-756e-4be8-b863-368986fbcf0a\") " pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.095829 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9393c0c7-756e-4be8-b863-368986fbcf0a-config\") pod \"route-controller-manager-695bbbd669-m87vv\" (UID: \"9393c0c7-756e-4be8-b863-368986fbcf0a\") " pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.095866 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.095876 4777 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.095885 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.095894 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbjcc\" (UniqueName: \"kubernetes.io/projected/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179-kube-api-access-fbjcc\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.096237 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1407cd97-3ca4-4f29-a6be-700428acce49" (UID: "1407cd97-3ca4-4f29-a6be-700428acce49"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.096258 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-client-ca" (OuterVolumeSpecName: "client-ca") pod "1407cd97-3ca4-4f29-a6be-700428acce49" (UID: "1407cd97-3ca4-4f29-a6be-700428acce49"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.096400 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-config" (OuterVolumeSpecName: "config") pod "1407cd97-3ca4-4f29-a6be-700428acce49" (UID: "1407cd97-3ca4-4f29-a6be-700428acce49"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.096843 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9393c0c7-756e-4be8-b863-368986fbcf0a-client-ca\") pod \"route-controller-manager-695bbbd669-m87vv\" (UID: \"9393c0c7-756e-4be8-b863-368986fbcf0a\") " pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.097007 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9393c0c7-756e-4be8-b863-368986fbcf0a-config\") pod \"route-controller-manager-695bbbd669-m87vv\" (UID: \"9393c0c7-756e-4be8-b863-368986fbcf0a\") " pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.098015 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1407cd97-3ca4-4f29-a6be-700428acce49-kube-api-access-gqtjb" (OuterVolumeSpecName: "kube-api-access-gqtjb") pod "1407cd97-3ca4-4f29-a6be-700428acce49" (UID: "1407cd97-3ca4-4f29-a6be-700428acce49"). InnerVolumeSpecName "kube-api-access-gqtjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.098117 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1407cd97-3ca4-4f29-a6be-700428acce49-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1407cd97-3ca4-4f29-a6be-700428acce49" (UID: "1407cd97-3ca4-4f29-a6be-700428acce49"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.099782 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9393c0c7-756e-4be8-b863-368986fbcf0a-serving-cert\") pod \"route-controller-manager-695bbbd669-m87vv\" (UID: \"9393c0c7-756e-4be8-b863-368986fbcf0a\") " pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.112019 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpw45\" (UniqueName: \"kubernetes.io/projected/9393c0c7-756e-4be8-b863-368986fbcf0a-kube-api-access-dpw45\") pod \"route-controller-manager-695bbbd669-m87vv\" (UID: \"9393c0c7-756e-4be8-b863-368986fbcf0a\") " pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.196742 4777 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.196770 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.196781 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqtjb\" (UniqueName: \"kubernetes.io/projected/1407cd97-3ca4-4f29-a6be-700428acce49-kube-api-access-gqtjb\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.196791 4777 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1407cd97-3ca4-4f29-a6be-700428acce49-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.196798 4777 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1407cd97-3ca4-4f29-a6be-700428acce49-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.231084 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.488673 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv"] Nov 24 17:13:36 crc kubenswrapper[4777]: W1124 17:13:36.495534 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9393c0c7_756e_4be8_b863_368986fbcf0a.slice/crio-7c0b67eee0cd8f5ed230c626cb4edca2749215f383cbd5cb5430208c74b0feac WatchSource:0}: Error finding container 7c0b67eee0cd8f5ed230c626cb4edca2749215f383cbd5cb5430208c74b0feac: Status 404 returned error can't find the container with id 7c0b67eee0cd8f5ed230c626cb4edca2749215f383cbd5cb5430208c74b0feac Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.500312 4777 generic.go:334] "Generic (PLEG): container finished" podID="bfa08af0-0c1f-446a-9a55-4aeb0cdb7179" containerID="edaccbc37f80bda8ea2f651d7a25c1e889c3136277f3b9d4c9515022c2de66b9" exitCode=0 Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.500350 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" event={"ID":"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179","Type":"ContainerDied","Data":"edaccbc37f80bda8ea2f651d7a25c1e889c3136277f3b9d4c9515022c2de66b9"} Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.500401 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.500429 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d" event={"ID":"bfa08af0-0c1f-446a-9a55-4aeb0cdb7179","Type":"ContainerDied","Data":"c42ce4aca5287466cd92cb6a62d0c7922c999ee4c7e1a2ce5049264e60dbfbb7"} Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.500452 4777 scope.go:117] "RemoveContainer" containerID="edaccbc37f80bda8ea2f651d7a25c1e889c3136277f3b9d4c9515022c2de66b9" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.507018 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-8wtx5" event={"ID":"3ffb9c2f-e94c-4fda-8c82-a34c045f9435","Type":"ContainerStarted","Data":"0fda5882d834c9db3c71c94ec25ee24771cf89f8c8474df8d52cb3573009c59b"} Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.509549 4777 generic.go:334] "Generic (PLEG): container finished" podID="1407cd97-3ca4-4f29-a6be-700428acce49" containerID="e6e5c1264bc703abcb3f3ee75b51fcdd59b7ce26df13e617ec5ca9d372043b52" exitCode=0 Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.509579 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" event={"ID":"1407cd97-3ca4-4f29-a6be-700428acce49","Type":"ContainerDied","Data":"e6e5c1264bc703abcb3f3ee75b51fcdd59b7ce26df13e617ec5ca9d372043b52"} Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.509595 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" event={"ID":"1407cd97-3ca4-4f29-a6be-700428acce49","Type":"ContainerDied","Data":"dae93807c0e49a974fb2149a3e0bb89a10382cfef96c8d6ae853fdc368e06ad5"} Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.509657 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f579785f-bddj2" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.525615 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-8wtx5" podStartSLOduration=2.036134029 podStartE2EDuration="4.525596851s" podCreationTimestamp="2025-11-24 17:13:32 +0000 UTC" firstStartedPulling="2025-11-24 17:13:33.134679702 +0000 UTC m=+771.293874751" lastFinishedPulling="2025-11-24 17:13:35.624142504 +0000 UTC m=+773.783337573" observedRunningTime="2025-11-24 17:13:36.52486625 +0000 UTC m=+774.684061309" watchObservedRunningTime="2025-11-24 17:13:36.525596851 +0000 UTC m=+774.684791900" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.537220 4777 scope.go:117] "RemoveContainer" containerID="edaccbc37f80bda8ea2f651d7a25c1e889c3136277f3b9d4c9515022c2de66b9" Nov 24 17:13:36 crc kubenswrapper[4777]: E1124 17:13:36.537656 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edaccbc37f80bda8ea2f651d7a25c1e889c3136277f3b9d4c9515022c2de66b9\": container with ID starting with edaccbc37f80bda8ea2f651d7a25c1e889c3136277f3b9d4c9515022c2de66b9 not found: ID does not exist" containerID="edaccbc37f80bda8ea2f651d7a25c1e889c3136277f3b9d4c9515022c2de66b9" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.537697 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edaccbc37f80bda8ea2f651d7a25c1e889c3136277f3b9d4c9515022c2de66b9"} err="failed to get container status \"edaccbc37f80bda8ea2f651d7a25c1e889c3136277f3b9d4c9515022c2de66b9\": rpc error: code = NotFound desc = could not find container \"edaccbc37f80bda8ea2f651d7a25c1e889c3136277f3b9d4c9515022c2de66b9\": container with ID starting with edaccbc37f80bda8ea2f651d7a25c1e889c3136277f3b9d4c9515022c2de66b9 not found: ID does not exist" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.537780 4777 scope.go:117] "RemoveContainer" containerID="e6e5c1264bc703abcb3f3ee75b51fcdd59b7ce26df13e617ec5ca9d372043b52" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.559087 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d"] Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.563304 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5548cdb5fc-f9v4d"] Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.567463 4777 scope.go:117] "RemoveContainer" containerID="e6e5c1264bc703abcb3f3ee75b51fcdd59b7ce26df13e617ec5ca9d372043b52" Nov 24 17:13:36 crc kubenswrapper[4777]: E1124 17:13:36.568156 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6e5c1264bc703abcb3f3ee75b51fcdd59b7ce26df13e617ec5ca9d372043b52\": container with ID starting with e6e5c1264bc703abcb3f3ee75b51fcdd59b7ce26df13e617ec5ca9d372043b52 not found: ID does not exist" containerID="e6e5c1264bc703abcb3f3ee75b51fcdd59b7ce26df13e617ec5ca9d372043b52" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.568191 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6e5c1264bc703abcb3f3ee75b51fcdd59b7ce26df13e617ec5ca9d372043b52"} err="failed to get container status \"e6e5c1264bc703abcb3f3ee75b51fcdd59b7ce26df13e617ec5ca9d372043b52\": rpc error: code = NotFound desc = could not find container \"e6e5c1264bc703abcb3f3ee75b51fcdd59b7ce26df13e617ec5ca9d372043b52\": container with ID starting with e6e5c1264bc703abcb3f3ee75b51fcdd59b7ce26df13e617ec5ca9d372043b52 not found: ID does not exist" Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.568850 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5f579785f-bddj2"] Nov 24 17:13:36 crc kubenswrapper[4777]: I1124 17:13:36.571747 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5f579785f-bddj2"] Nov 24 17:13:37 crc kubenswrapper[4777]: I1124 17:13:37.255805 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1407cd97-3ca4-4f29-a6be-700428acce49" path="/var/lib/kubelet/pods/1407cd97-3ca4-4f29-a6be-700428acce49/volumes" Nov 24 17:13:37 crc kubenswrapper[4777]: I1124 17:13:37.256728 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfa08af0-0c1f-446a-9a55-4aeb0cdb7179" path="/var/lib/kubelet/pods/bfa08af0-0c1f-446a-9a55-4aeb0cdb7179/volumes" Nov 24 17:13:37 crc kubenswrapper[4777]: I1124 17:13:37.526777 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" event={"ID":"9393c0c7-756e-4be8-b863-368986fbcf0a","Type":"ContainerStarted","Data":"e5c6a1a5046699c6cf5bb6d887b87a8d6952d11c52e9f30014c7d24011180b6e"} Nov 24 17:13:37 crc kubenswrapper[4777]: I1124 17:13:37.527049 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" event={"ID":"9393c0c7-756e-4be8-b863-368986fbcf0a","Type":"ContainerStarted","Data":"7c0b67eee0cd8f5ed230c626cb4edca2749215f383cbd5cb5430208c74b0feac"} Nov 24 17:13:37 crc kubenswrapper[4777]: I1124 17:13:37.527946 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:37 crc kubenswrapper[4777]: I1124 17:13:37.531803 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" Nov 24 17:13:37 crc kubenswrapper[4777]: I1124 17:13:37.541008 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-695bbbd669-m87vv" podStartSLOduration=3.540988966 podStartE2EDuration="3.540988966s" podCreationTimestamp="2025-11-24 17:13:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:13:37.53836154 +0000 UTC m=+775.697556579" watchObservedRunningTime="2025-11-24 17:13:37.540988966 +0000 UTC m=+775.700184015" Nov 24 17:13:37 crc kubenswrapper[4777]: I1124 17:13:37.999270 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq"] Nov 24 17:13:37 crc kubenswrapper[4777]: E1124 17:13:37.999702 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1407cd97-3ca4-4f29-a6be-700428acce49" containerName="controller-manager" Nov 24 17:13:37 crc kubenswrapper[4777]: I1124 17:13:37.999803 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="1407cd97-3ca4-4f29-a6be-700428acce49" containerName="controller-manager" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.000056 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="1407cd97-3ca4-4f29-a6be-700428acce49" containerName="controller-manager" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.000597 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.004119 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.004466 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.006003 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.006239 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.006488 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.012559 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.014419 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.021800 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq"] Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.117997 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxrrp\" (UniqueName: \"kubernetes.io/projected/e31e481f-8887-499b-ba2c-66bfc3554c89-kube-api-access-lxrrp\") pod \"controller-manager-ff5d79dd6-8wwrq\" (UID: \"e31e481f-8887-499b-ba2c-66bfc3554c89\") " pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.118311 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e31e481f-8887-499b-ba2c-66bfc3554c89-serving-cert\") pod \"controller-manager-ff5d79dd6-8wwrq\" (UID: \"e31e481f-8887-499b-ba2c-66bfc3554c89\") " pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.118421 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e31e481f-8887-499b-ba2c-66bfc3554c89-proxy-ca-bundles\") pod \"controller-manager-ff5d79dd6-8wwrq\" (UID: \"e31e481f-8887-499b-ba2c-66bfc3554c89\") " pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.118529 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e31e481f-8887-499b-ba2c-66bfc3554c89-client-ca\") pod \"controller-manager-ff5d79dd6-8wwrq\" (UID: \"e31e481f-8887-499b-ba2c-66bfc3554c89\") " pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.118653 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e31e481f-8887-499b-ba2c-66bfc3554c89-config\") pod \"controller-manager-ff5d79dd6-8wwrq\" (UID: \"e31e481f-8887-499b-ba2c-66bfc3554c89\") " pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.219844 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxrrp\" (UniqueName: \"kubernetes.io/projected/e31e481f-8887-499b-ba2c-66bfc3554c89-kube-api-access-lxrrp\") pod \"controller-manager-ff5d79dd6-8wwrq\" (UID: \"e31e481f-8887-499b-ba2c-66bfc3554c89\") " pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.219937 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e31e481f-8887-499b-ba2c-66bfc3554c89-serving-cert\") pod \"controller-manager-ff5d79dd6-8wwrq\" (UID: \"e31e481f-8887-499b-ba2c-66bfc3554c89\") " pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.220008 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e31e481f-8887-499b-ba2c-66bfc3554c89-proxy-ca-bundles\") pod \"controller-manager-ff5d79dd6-8wwrq\" (UID: \"e31e481f-8887-499b-ba2c-66bfc3554c89\") " pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.220059 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e31e481f-8887-499b-ba2c-66bfc3554c89-client-ca\") pod \"controller-manager-ff5d79dd6-8wwrq\" (UID: \"e31e481f-8887-499b-ba2c-66bfc3554c89\") " pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.220121 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e31e481f-8887-499b-ba2c-66bfc3554c89-config\") pod \"controller-manager-ff5d79dd6-8wwrq\" (UID: \"e31e481f-8887-499b-ba2c-66bfc3554c89\") " pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.221771 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e31e481f-8887-499b-ba2c-66bfc3554c89-client-ca\") pod \"controller-manager-ff5d79dd6-8wwrq\" (UID: \"e31e481f-8887-499b-ba2c-66bfc3554c89\") " pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.222605 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e31e481f-8887-499b-ba2c-66bfc3554c89-proxy-ca-bundles\") pod \"controller-manager-ff5d79dd6-8wwrq\" (UID: \"e31e481f-8887-499b-ba2c-66bfc3554c89\") " pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.222891 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e31e481f-8887-499b-ba2c-66bfc3554c89-config\") pod \"controller-manager-ff5d79dd6-8wwrq\" (UID: \"e31e481f-8887-499b-ba2c-66bfc3554c89\") " pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.230829 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e31e481f-8887-499b-ba2c-66bfc3554c89-serving-cert\") pod \"controller-manager-ff5d79dd6-8wwrq\" (UID: \"e31e481f-8887-499b-ba2c-66bfc3554c89\") " pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.248368 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxrrp\" (UniqueName: \"kubernetes.io/projected/e31e481f-8887-499b-ba2c-66bfc3554c89-kube-api-access-lxrrp\") pod \"controller-manager-ff5d79dd6-8wwrq\" (UID: \"e31e481f-8887-499b-ba2c-66bfc3554c89\") " pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.357643 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:38 crc kubenswrapper[4777]: I1124 17:13:38.852853 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq"] Nov 24 17:13:39 crc kubenswrapper[4777]: I1124 17:13:39.539163 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" event={"ID":"e31e481f-8887-499b-ba2c-66bfc3554c89","Type":"ContainerStarted","Data":"20e5051effb58674e3c43e2dc111e53ed955496e8f9278e29834f1460455cdb1"} Nov 24 17:13:39 crc kubenswrapper[4777]: I1124 17:13:39.539520 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" event={"ID":"e31e481f-8887-499b-ba2c-66bfc3554c89","Type":"ContainerStarted","Data":"fc68331376019a9c83aa570ad874cb966df8c17a76fc9019e19b180fc5c9410b"} Nov 24 17:13:39 crc kubenswrapper[4777]: I1124 17:13:39.561047 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" podStartSLOduration=5.561032543 podStartE2EDuration="5.561032543s" podCreationTimestamp="2025-11-24 17:13:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:13:39.55851074 +0000 UTC m=+777.717705789" watchObservedRunningTime="2025-11-24 17:13:39.561032543 +0000 UTC m=+777.720227592" Nov 24 17:13:40 crc kubenswrapper[4777]: I1124 17:13:40.545168 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:40 crc kubenswrapper[4777]: I1124 17:13:40.550867 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-ff5d79dd6-8wwrq" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.615019 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6"] Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.616536 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.618067 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-m7wgh"] Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.619757 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-m7wgh" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.626067 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.626372 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-fm7nc" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.631936 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-wnlkc"] Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.634010 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.650464 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6"] Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.653878 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-m7wgh"] Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.736934 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq"] Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.737622 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.739386 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.739603 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-6tg5b" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.739640 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.751825 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq"] Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.767304 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6aa57fcf-4602-4a14-8f93-d9881aad4c8d-ovs-socket\") pod \"nmstate-handler-wnlkc\" (UID: \"6aa57fcf-4602-4a14-8f93-d9881aad4c8d\") " pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.767566 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr2sg\" (UniqueName: \"kubernetes.io/projected/64c0b83e-a764-4494-b384-727c83c914a2-kube-api-access-dr2sg\") pod \"nmstate-webhook-6b89b748d8-z59z6\" (UID: \"64c0b83e-a764-4494-b384-727c83c914a2\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.767693 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6aa57fcf-4602-4a14-8f93-d9881aad4c8d-nmstate-lock\") pod \"nmstate-handler-wnlkc\" (UID: \"6aa57fcf-4602-4a14-8f93-d9881aad4c8d\") " pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.767812 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j98xq\" (UniqueName: \"kubernetes.io/projected/2cb2400c-9df2-4e5f-95e9-3e8d22b58d3c-kube-api-access-j98xq\") pod \"nmstate-metrics-5dcf9c57c5-m7wgh\" (UID: \"2cb2400c-9df2-4e5f-95e9-3e8d22b58d3c\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-m7wgh" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.767838 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/64c0b83e-a764-4494-b384-727c83c914a2-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-z59z6\" (UID: \"64c0b83e-a764-4494-b384-727c83c914a2\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.767872 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6aa57fcf-4602-4a14-8f93-d9881aad4c8d-dbus-socket\") pod \"nmstate-handler-wnlkc\" (UID: \"6aa57fcf-4602-4a14-8f93-d9881aad4c8d\") " pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.767918 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl5rr\" (UniqueName: \"kubernetes.io/projected/6aa57fcf-4602-4a14-8f93-d9881aad4c8d-kube-api-access-sl5rr\") pod \"nmstate-handler-wnlkc\" (UID: \"6aa57fcf-4602-4a14-8f93-d9881aad4c8d\") " pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.869335 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl5rr\" (UniqueName: \"kubernetes.io/projected/6aa57fcf-4602-4a14-8f93-d9881aad4c8d-kube-api-access-sl5rr\") pod \"nmstate-handler-wnlkc\" (UID: \"6aa57fcf-4602-4a14-8f93-d9881aad4c8d\") " pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.869400 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2f2d18e5-97fa-47f6-9194-760ff54fdddc-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-p6sjq\" (UID: \"2f2d18e5-97fa-47f6-9194-760ff54fdddc\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.869447 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6aa57fcf-4602-4a14-8f93-d9881aad4c8d-ovs-socket\") pod \"nmstate-handler-wnlkc\" (UID: \"6aa57fcf-4602-4a14-8f93-d9881aad4c8d\") " pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.869478 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr2sg\" (UniqueName: \"kubernetes.io/projected/64c0b83e-a764-4494-b384-727c83c914a2-kube-api-access-dr2sg\") pod \"nmstate-webhook-6b89b748d8-z59z6\" (UID: \"64c0b83e-a764-4494-b384-727c83c914a2\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.869502 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6aa57fcf-4602-4a14-8f93-d9881aad4c8d-nmstate-lock\") pod \"nmstate-handler-wnlkc\" (UID: \"6aa57fcf-4602-4a14-8f93-d9881aad4c8d\") " pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.869559 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j98xq\" (UniqueName: \"kubernetes.io/projected/2cb2400c-9df2-4e5f-95e9-3e8d22b58d3c-kube-api-access-j98xq\") pod \"nmstate-metrics-5dcf9c57c5-m7wgh\" (UID: \"2cb2400c-9df2-4e5f-95e9-3e8d22b58d3c\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-m7wgh" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.869583 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/64c0b83e-a764-4494-b384-727c83c914a2-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-z59z6\" (UID: \"64c0b83e-a764-4494-b384-727c83c914a2\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.869611 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6aa57fcf-4602-4a14-8f93-d9881aad4c8d-dbus-socket\") pod \"nmstate-handler-wnlkc\" (UID: \"6aa57fcf-4602-4a14-8f93-d9881aad4c8d\") " pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.869635 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2f2d18e5-97fa-47f6-9194-760ff54fdddc-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-p6sjq\" (UID: \"2f2d18e5-97fa-47f6-9194-760ff54fdddc\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.869656 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nw5w\" (UniqueName: \"kubernetes.io/projected/2f2d18e5-97fa-47f6-9194-760ff54fdddc-kube-api-access-9nw5w\") pod \"nmstate-console-plugin-5874bd7bc5-p6sjq\" (UID: \"2f2d18e5-97fa-47f6-9194-760ff54fdddc\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.870062 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6aa57fcf-4602-4a14-8f93-d9881aad4c8d-ovs-socket\") pod \"nmstate-handler-wnlkc\" (UID: \"6aa57fcf-4602-4a14-8f93-d9881aad4c8d\") " pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.870251 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6aa57fcf-4602-4a14-8f93-d9881aad4c8d-nmstate-lock\") pod \"nmstate-handler-wnlkc\" (UID: \"6aa57fcf-4602-4a14-8f93-d9881aad4c8d\") " pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.870580 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6aa57fcf-4602-4a14-8f93-d9881aad4c8d-dbus-socket\") pod \"nmstate-handler-wnlkc\" (UID: \"6aa57fcf-4602-4a14-8f93-d9881aad4c8d\") " pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.876072 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/64c0b83e-a764-4494-b384-727c83c914a2-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-z59z6\" (UID: \"64c0b83e-a764-4494-b384-727c83c914a2\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.889428 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl5rr\" (UniqueName: \"kubernetes.io/projected/6aa57fcf-4602-4a14-8f93-d9881aad4c8d-kube-api-access-sl5rr\") pod \"nmstate-handler-wnlkc\" (UID: \"6aa57fcf-4602-4a14-8f93-d9881aad4c8d\") " pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.889897 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr2sg\" (UniqueName: \"kubernetes.io/projected/64c0b83e-a764-4494-b384-727c83c914a2-kube-api-access-dr2sg\") pod \"nmstate-webhook-6b89b748d8-z59z6\" (UID: \"64c0b83e-a764-4494-b384-727c83c914a2\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.893676 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j98xq\" (UniqueName: \"kubernetes.io/projected/2cb2400c-9df2-4e5f-95e9-3e8d22b58d3c-kube-api-access-j98xq\") pod \"nmstate-metrics-5dcf9c57c5-m7wgh\" (UID: \"2cb2400c-9df2-4e5f-95e9-3e8d22b58d3c\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-m7wgh" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.931599 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6b8f557976-8xg89"] Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.932304 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.947919 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.959837 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-m7wgh" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.961425 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6b8f557976-8xg89"] Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.970820 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2f2d18e5-97fa-47f6-9194-760ff54fdddc-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-p6sjq\" (UID: \"2f2d18e5-97fa-47f6-9194-760ff54fdddc\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.970903 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nw5w\" (UniqueName: \"kubernetes.io/projected/2f2d18e5-97fa-47f6-9194-760ff54fdddc-kube-api-access-9nw5w\") pod \"nmstate-console-plugin-5874bd7bc5-p6sjq\" (UID: \"2f2d18e5-97fa-47f6-9194-760ff54fdddc\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.970921 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2f2d18e5-97fa-47f6-9194-760ff54fdddc-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-p6sjq\" (UID: \"2f2d18e5-97fa-47f6-9194-760ff54fdddc\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.971521 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.972192 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2f2d18e5-97fa-47f6-9194-760ff54fdddc-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-p6sjq\" (UID: \"2f2d18e5-97fa-47f6-9194-760ff54fdddc\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq" Nov 24 17:13:41 crc kubenswrapper[4777]: I1124 17:13:41.986329 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2f2d18e5-97fa-47f6-9194-760ff54fdddc-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-p6sjq\" (UID: \"2f2d18e5-97fa-47f6-9194-760ff54fdddc\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.003535 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nw5w\" (UniqueName: \"kubernetes.io/projected/2f2d18e5-97fa-47f6-9194-760ff54fdddc-kube-api-access-9nw5w\") pod \"nmstate-console-plugin-5874bd7bc5-p6sjq\" (UID: \"2f2d18e5-97fa-47f6-9194-760ff54fdddc\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq" Nov 24 17:13:42 crc kubenswrapper[4777]: W1124 17:13:42.010404 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6aa57fcf_4602_4a14_8f93_d9881aad4c8d.slice/crio-3e6a34abc05a8e032aeaac8c6e96d095af07669ce2fa46de7c4eef8a819403aa WatchSource:0}: Error finding container 3e6a34abc05a8e032aeaac8c6e96d095af07669ce2fa46de7c4eef8a819403aa: Status 404 returned error can't find the container with id 3e6a34abc05a8e032aeaac8c6e96d095af07669ce2fa46de7c4eef8a819403aa Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.052423 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.075354 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a46e0ce2-add0-456d-bff5-5001878b273e-console-config\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.075414 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a46e0ce2-add0-456d-bff5-5001878b273e-trusted-ca-bundle\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.075467 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a46e0ce2-add0-456d-bff5-5001878b273e-oauth-serving-cert\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.075515 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a46e0ce2-add0-456d-bff5-5001878b273e-service-ca\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.075605 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqvq2\" (UniqueName: \"kubernetes.io/projected/a46e0ce2-add0-456d-bff5-5001878b273e-kube-api-access-lqvq2\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.075703 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a46e0ce2-add0-456d-bff5-5001878b273e-console-oauth-config\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.075731 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a46e0ce2-add0-456d-bff5-5001878b273e-console-serving-cert\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.177320 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqvq2\" (UniqueName: \"kubernetes.io/projected/a46e0ce2-add0-456d-bff5-5001878b273e-kube-api-access-lqvq2\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.177369 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a46e0ce2-add0-456d-bff5-5001878b273e-console-oauth-config\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.177392 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a46e0ce2-add0-456d-bff5-5001878b273e-console-serving-cert\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.177416 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a46e0ce2-add0-456d-bff5-5001878b273e-console-config\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.177435 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a46e0ce2-add0-456d-bff5-5001878b273e-trusted-ca-bundle\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.177463 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a46e0ce2-add0-456d-bff5-5001878b273e-oauth-serving-cert\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.177488 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a46e0ce2-add0-456d-bff5-5001878b273e-service-ca\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.178582 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a46e0ce2-add0-456d-bff5-5001878b273e-oauth-serving-cert\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.178686 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a46e0ce2-add0-456d-bff5-5001878b273e-trusted-ca-bundle\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.178799 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a46e0ce2-add0-456d-bff5-5001878b273e-service-ca\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.179074 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a46e0ce2-add0-456d-bff5-5001878b273e-console-config\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.183678 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a46e0ce2-add0-456d-bff5-5001878b273e-console-oauth-config\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.183823 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a46e0ce2-add0-456d-bff5-5001878b273e-console-serving-cert\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.191410 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqvq2\" (UniqueName: \"kubernetes.io/projected/a46e0ce2-add0-456d-bff5-5001878b273e-kube-api-access-lqvq2\") pod \"console-6b8f557976-8xg89\" (UID: \"a46e0ce2-add0-456d-bff5-5001878b273e\") " pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.251827 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.384806 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-m7wgh"] Nov 24 17:13:42 crc kubenswrapper[4777]: W1124 17:13:42.390750 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cb2400c_9df2_4e5f_95e9_3e8d22b58d3c.slice/crio-fcbe793af68ed9f116c8b76941e3c15ebc676d9164f919bc740d982bc1b8348b WatchSource:0}: Error finding container fcbe793af68ed9f116c8b76941e3c15ebc676d9164f919bc740d982bc1b8348b: Status 404 returned error can't find the container with id fcbe793af68ed9f116c8b76941e3c15ebc676d9164f919bc740d982bc1b8348b Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.439153 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6"] Nov 24 17:13:42 crc kubenswrapper[4777]: W1124 17:13:42.446824 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64c0b83e_a764_4494_b384_727c83c914a2.slice/crio-2ce073766e5234057df06ee7b413c9b5929ec822904065422791017890dce8e8 WatchSource:0}: Error finding container 2ce073766e5234057df06ee7b413c9b5929ec822904065422791017890dce8e8: Status 404 returned error can't find the container with id 2ce073766e5234057df06ee7b413c9b5929ec822904065422791017890dce8e8 Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.480608 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq"] Nov 24 17:13:42 crc kubenswrapper[4777]: W1124 17:13:42.490877 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f2d18e5_97fa_47f6_9194_760ff54fdddc.slice/crio-21cb906f6cd136bbf1fad9ec29767aed0bfc7d62c8ba17c2380949b5da523c1b WatchSource:0}: Error finding container 21cb906f6cd136bbf1fad9ec29767aed0bfc7d62c8ba17c2380949b5da523c1b: Status 404 returned error can't find the container with id 21cb906f6cd136bbf1fad9ec29767aed0bfc7d62c8ba17c2380949b5da523c1b Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.556111 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6" event={"ID":"64c0b83e-a764-4494-b384-727c83c914a2","Type":"ContainerStarted","Data":"2ce073766e5234057df06ee7b413c9b5929ec822904065422791017890dce8e8"} Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.557909 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-wnlkc" event={"ID":"6aa57fcf-4602-4a14-8f93-d9881aad4c8d","Type":"ContainerStarted","Data":"3e6a34abc05a8e032aeaac8c6e96d095af07669ce2fa46de7c4eef8a819403aa"} Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.558732 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-m7wgh" event={"ID":"2cb2400c-9df2-4e5f-95e9-3e8d22b58d3c","Type":"ContainerStarted","Data":"fcbe793af68ed9f116c8b76941e3c15ebc676d9164f919bc740d982bc1b8348b"} Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.560163 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq" event={"ID":"2f2d18e5-97fa-47f6-9194-760ff54fdddc","Type":"ContainerStarted","Data":"21cb906f6cd136bbf1fad9ec29767aed0bfc7d62c8ba17c2380949b5da523c1b"} Nov 24 17:13:42 crc kubenswrapper[4777]: I1124 17:13:42.688116 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6b8f557976-8xg89"] Nov 24 17:13:42 crc kubenswrapper[4777]: W1124 17:13:42.697633 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda46e0ce2_add0_456d_bff5_5001878b273e.slice/crio-6597f12d337eff789301ac0d6988cdb8401e8f90530937849f4ce3f9a47f3f61 WatchSource:0}: Error finding container 6597f12d337eff789301ac0d6988cdb8401e8f90530937849f4ce3f9a47f3f61: Status 404 returned error can't find the container with id 6597f12d337eff789301ac0d6988cdb8401e8f90530937849f4ce3f9a47f3f61 Nov 24 17:13:43 crc kubenswrapper[4777]: I1124 17:13:43.101280 4777 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 17:13:43 crc kubenswrapper[4777]: I1124 17:13:43.567764 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6b8f557976-8xg89" event={"ID":"a46e0ce2-add0-456d-bff5-5001878b273e","Type":"ContainerStarted","Data":"c79ad93cdf29bcac4084c5720587414d2b46a4752b2d15aeacd64574059d55df"} Nov 24 17:13:43 crc kubenswrapper[4777]: I1124 17:13:43.567813 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6b8f557976-8xg89" event={"ID":"a46e0ce2-add0-456d-bff5-5001878b273e","Type":"ContainerStarted","Data":"6597f12d337eff789301ac0d6988cdb8401e8f90530937849f4ce3f9a47f3f61"} Nov 24 17:13:43 crc kubenswrapper[4777]: I1124 17:13:43.584106 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6b8f557976-8xg89" podStartSLOduration=2.584082647 podStartE2EDuration="2.584082647s" podCreationTimestamp="2025-11-24 17:13:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:13:43.583055077 +0000 UTC m=+781.742250136" watchObservedRunningTime="2025-11-24 17:13:43.584082647 +0000 UTC m=+781.743277716" Nov 24 17:13:45 crc kubenswrapper[4777]: I1124 17:13:45.583937 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq" event={"ID":"2f2d18e5-97fa-47f6-9194-760ff54fdddc","Type":"ContainerStarted","Data":"2e5b711bc100d18ff09e804c4ad5e869855ae40c03960b9160fc2eb3c12ee8ef"} Nov 24 17:13:45 crc kubenswrapper[4777]: I1124 17:13:45.588869 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6" event={"ID":"64c0b83e-a764-4494-b384-727c83c914a2","Type":"ContainerStarted","Data":"a780795d181a9f4628ada710bcd0f1dcb56e89a682cd728c1c21014937c4f701"} Nov 24 17:13:45 crc kubenswrapper[4777]: I1124 17:13:45.588995 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6" Nov 24 17:13:45 crc kubenswrapper[4777]: I1124 17:13:45.591947 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-wnlkc" event={"ID":"6aa57fcf-4602-4a14-8f93-d9881aad4c8d","Type":"ContainerStarted","Data":"a72e7986b52896ca39c5d878ffb4858e0e60cf4097abc33acd5fa334d2b6e51b"} Nov 24 17:13:45 crc kubenswrapper[4777]: I1124 17:13:45.592024 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:45 crc kubenswrapper[4777]: I1124 17:13:45.593792 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-m7wgh" event={"ID":"2cb2400c-9df2-4e5f-95e9-3e8d22b58d3c","Type":"ContainerStarted","Data":"d65acecd65c502eac126420f99ee31912003152c0a1b15efad80d5d96687d2a2"} Nov 24 17:13:45 crc kubenswrapper[4777]: I1124 17:13:45.600624 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p6sjq" podStartSLOduration=2.139256547 podStartE2EDuration="4.600606349s" podCreationTimestamp="2025-11-24 17:13:41 +0000 UTC" firstStartedPulling="2025-11-24 17:13:42.493431542 +0000 UTC m=+780.652626591" lastFinishedPulling="2025-11-24 17:13:44.954781334 +0000 UTC m=+783.113976393" observedRunningTime="2025-11-24 17:13:45.599451026 +0000 UTC m=+783.758646105" watchObservedRunningTime="2025-11-24 17:13:45.600606349 +0000 UTC m=+783.759801408" Nov 24 17:13:45 crc kubenswrapper[4777]: I1124 17:13:45.628396 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6" podStartSLOduration=2.087898791 podStartE2EDuration="4.628369932s" podCreationTimestamp="2025-11-24 17:13:41 +0000 UTC" firstStartedPulling="2025-11-24 17:13:42.449254554 +0000 UTC m=+780.608449603" lastFinishedPulling="2025-11-24 17:13:44.989725695 +0000 UTC m=+783.148920744" observedRunningTime="2025-11-24 17:13:45.621887165 +0000 UTC m=+783.781082244" watchObservedRunningTime="2025-11-24 17:13:45.628369932 +0000 UTC m=+783.787565021" Nov 24 17:13:45 crc kubenswrapper[4777]: I1124 17:13:45.644921 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-wnlkc" podStartSLOduration=1.703844331 podStartE2EDuration="4.644901421s" podCreationTimestamp="2025-11-24 17:13:41 +0000 UTC" firstStartedPulling="2025-11-24 17:13:42.012158419 +0000 UTC m=+780.171353478" lastFinishedPulling="2025-11-24 17:13:44.953215479 +0000 UTC m=+783.112410568" observedRunningTime="2025-11-24 17:13:45.643605083 +0000 UTC m=+783.802800142" watchObservedRunningTime="2025-11-24 17:13:45.644901421 +0000 UTC m=+783.804096470" Nov 24 17:13:47 crc kubenswrapper[4777]: I1124 17:13:47.611208 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-m7wgh" event={"ID":"2cb2400c-9df2-4e5f-95e9-3e8d22b58d3c","Type":"ContainerStarted","Data":"d5910214b2b74d2a0c8ec2f64fb0f0802771a0584da1bc65f8f8844511491d01"} Nov 24 17:13:47 crc kubenswrapper[4777]: I1124 17:13:47.639343 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-m7wgh" podStartSLOduration=2.030135101 podStartE2EDuration="6.639315892s" podCreationTimestamp="2025-11-24 17:13:41 +0000 UTC" firstStartedPulling="2025-11-24 17:13:42.393067579 +0000 UTC m=+780.552262628" lastFinishedPulling="2025-11-24 17:13:47.00224836 +0000 UTC m=+785.161443419" observedRunningTime="2025-11-24 17:13:47.638073826 +0000 UTC m=+785.797268915" watchObservedRunningTime="2025-11-24 17:13:47.639315892 +0000 UTC m=+785.798510981" Nov 24 17:13:52 crc kubenswrapper[4777]: I1124 17:13:52.002731 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-wnlkc" Nov 24 17:13:52 crc kubenswrapper[4777]: I1124 17:13:52.253378 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:52 crc kubenswrapper[4777]: I1124 17:13:52.253796 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:52 crc kubenswrapper[4777]: I1124 17:13:52.262120 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:52 crc kubenswrapper[4777]: I1124 17:13:52.663308 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6b8f557976-8xg89" Nov 24 17:13:52 crc kubenswrapper[4777]: I1124 17:13:52.741376 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-jxbj9"] Nov 24 17:14:01 crc kubenswrapper[4777]: I1124 17:14:01.954809 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-z59z6" Nov 24 17:14:07 crc kubenswrapper[4777]: I1124 17:14:07.610318 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9b95k"] Nov 24 17:14:07 crc kubenswrapper[4777]: I1124 17:14:07.613071 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:07 crc kubenswrapper[4777]: I1124 17:14:07.619045 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9b95k"] Nov 24 17:14:07 crc kubenswrapper[4777]: I1124 17:14:07.649644 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4670321e-4241-4686-8eaa-860bac08a44f-catalog-content\") pod \"redhat-operators-9b95k\" (UID: \"4670321e-4241-4686-8eaa-860bac08a44f\") " pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:07 crc kubenswrapper[4777]: I1124 17:14:07.649737 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmpng\" (UniqueName: \"kubernetes.io/projected/4670321e-4241-4686-8eaa-860bac08a44f-kube-api-access-vmpng\") pod \"redhat-operators-9b95k\" (UID: \"4670321e-4241-4686-8eaa-860bac08a44f\") " pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:07 crc kubenswrapper[4777]: I1124 17:14:07.649817 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4670321e-4241-4686-8eaa-860bac08a44f-utilities\") pod \"redhat-operators-9b95k\" (UID: \"4670321e-4241-4686-8eaa-860bac08a44f\") " pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:07 crc kubenswrapper[4777]: I1124 17:14:07.750804 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4670321e-4241-4686-8eaa-860bac08a44f-utilities\") pod \"redhat-operators-9b95k\" (UID: \"4670321e-4241-4686-8eaa-860bac08a44f\") " pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:07 crc kubenswrapper[4777]: I1124 17:14:07.750866 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4670321e-4241-4686-8eaa-860bac08a44f-catalog-content\") pod \"redhat-operators-9b95k\" (UID: \"4670321e-4241-4686-8eaa-860bac08a44f\") " pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:07 crc kubenswrapper[4777]: I1124 17:14:07.750926 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmpng\" (UniqueName: \"kubernetes.io/projected/4670321e-4241-4686-8eaa-860bac08a44f-kube-api-access-vmpng\") pod \"redhat-operators-9b95k\" (UID: \"4670321e-4241-4686-8eaa-860bac08a44f\") " pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:07 crc kubenswrapper[4777]: I1124 17:14:07.751623 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4670321e-4241-4686-8eaa-860bac08a44f-utilities\") pod \"redhat-operators-9b95k\" (UID: \"4670321e-4241-4686-8eaa-860bac08a44f\") " pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:07 crc kubenswrapper[4777]: I1124 17:14:07.751702 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4670321e-4241-4686-8eaa-860bac08a44f-catalog-content\") pod \"redhat-operators-9b95k\" (UID: \"4670321e-4241-4686-8eaa-860bac08a44f\") " pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:07 crc kubenswrapper[4777]: I1124 17:14:07.777317 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmpng\" (UniqueName: \"kubernetes.io/projected/4670321e-4241-4686-8eaa-860bac08a44f-kube-api-access-vmpng\") pod \"redhat-operators-9b95k\" (UID: \"4670321e-4241-4686-8eaa-860bac08a44f\") " pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:07 crc kubenswrapper[4777]: I1124 17:14:07.936343 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:08 crc kubenswrapper[4777]: I1124 17:14:08.377360 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9b95k"] Nov 24 17:14:08 crc kubenswrapper[4777]: I1124 17:14:08.784144 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9b95k" event={"ID":"4670321e-4241-4686-8eaa-860bac08a44f","Type":"ContainerStarted","Data":"1b2fc694064baa7c4c4173e3df3122a28a3a224863e55d7246d4bff6f9afe0c9"} Nov 24 17:14:08 crc kubenswrapper[4777]: I1124 17:14:08.784489 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9b95k" event={"ID":"4670321e-4241-4686-8eaa-860bac08a44f","Type":"ContainerStarted","Data":"aeb3796be539fcbbebb4163fc6b7afe50ec6145e2800de1919d39591986947b2"} Nov 24 17:14:09 crc kubenswrapper[4777]: I1124 17:14:09.794929 4777 generic.go:334] "Generic (PLEG): container finished" podID="4670321e-4241-4686-8eaa-860bac08a44f" containerID="1b2fc694064baa7c4c4173e3df3122a28a3a224863e55d7246d4bff6f9afe0c9" exitCode=0 Nov 24 17:14:09 crc kubenswrapper[4777]: I1124 17:14:09.795035 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9b95k" event={"ID":"4670321e-4241-4686-8eaa-860bac08a44f","Type":"ContainerDied","Data":"1b2fc694064baa7c4c4173e3df3122a28a3a224863e55d7246d4bff6f9afe0c9"} Nov 24 17:14:12 crc kubenswrapper[4777]: I1124 17:14:12.838256 4777 generic.go:334] "Generic (PLEG): container finished" podID="4670321e-4241-4686-8eaa-860bac08a44f" containerID="dbd3b277144c52f170d546ba6ae621228c8b02c6a3aea9599c431a16f7573073" exitCode=0 Nov 24 17:14:12 crc kubenswrapper[4777]: I1124 17:14:12.838509 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9b95k" event={"ID":"4670321e-4241-4686-8eaa-860bac08a44f","Type":"ContainerDied","Data":"dbd3b277144c52f170d546ba6ae621228c8b02c6a3aea9599c431a16f7573073"} Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.195820 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4brxh"] Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.199305 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.213105 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4brxh"] Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.269940 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0952d762-98cc-4c67-959a-772d127f6eb2-catalog-content\") pod \"certified-operators-4brxh\" (UID: \"0952d762-98cc-4c67-959a-772d127f6eb2\") " pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.270007 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0952d762-98cc-4c67-959a-772d127f6eb2-utilities\") pod \"certified-operators-4brxh\" (UID: \"0952d762-98cc-4c67-959a-772d127f6eb2\") " pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.270116 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cj89\" (UniqueName: \"kubernetes.io/projected/0952d762-98cc-4c67-959a-772d127f6eb2-kube-api-access-5cj89\") pod \"certified-operators-4brxh\" (UID: \"0952d762-98cc-4c67-959a-772d127f6eb2\") " pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.370674 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cj89\" (UniqueName: \"kubernetes.io/projected/0952d762-98cc-4c67-959a-772d127f6eb2-kube-api-access-5cj89\") pod \"certified-operators-4brxh\" (UID: \"0952d762-98cc-4c67-959a-772d127f6eb2\") " pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.370739 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0952d762-98cc-4c67-959a-772d127f6eb2-catalog-content\") pod \"certified-operators-4brxh\" (UID: \"0952d762-98cc-4c67-959a-772d127f6eb2\") " pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.370765 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0952d762-98cc-4c67-959a-772d127f6eb2-utilities\") pod \"certified-operators-4brxh\" (UID: \"0952d762-98cc-4c67-959a-772d127f6eb2\") " pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.371218 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0952d762-98cc-4c67-959a-772d127f6eb2-utilities\") pod \"certified-operators-4brxh\" (UID: \"0952d762-98cc-4c67-959a-772d127f6eb2\") " pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.371385 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0952d762-98cc-4c67-959a-772d127f6eb2-catalog-content\") pod \"certified-operators-4brxh\" (UID: \"0952d762-98cc-4c67-959a-772d127f6eb2\") " pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.394100 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cj89\" (UniqueName: \"kubernetes.io/projected/0952d762-98cc-4c67-959a-772d127f6eb2-kube-api-access-5cj89\") pod \"certified-operators-4brxh\" (UID: \"0952d762-98cc-4c67-959a-772d127f6eb2\") " pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.532599 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.865389 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9b95k" event={"ID":"4670321e-4241-4686-8eaa-860bac08a44f","Type":"ContainerStarted","Data":"0a51dbe85c1afef6b369cac0aaab762498eb34832bce9b55d06a0f5b8757732c"} Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.883699 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9b95k" podStartSLOduration=3.306363697 podStartE2EDuration="7.883686247s" podCreationTimestamp="2025-11-24 17:14:07 +0000 UTC" firstStartedPulling="2025-11-24 17:14:09.80078299 +0000 UTC m=+807.959978039" lastFinishedPulling="2025-11-24 17:14:14.37810554 +0000 UTC m=+812.537300589" observedRunningTime="2025-11-24 17:14:14.880042621 +0000 UTC m=+813.039237670" watchObservedRunningTime="2025-11-24 17:14:14.883686247 +0000 UTC m=+813.042881296" Nov 24 17:14:14 crc kubenswrapper[4777]: I1124 17:14:14.971317 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4brxh"] Nov 24 17:14:15 crc kubenswrapper[4777]: I1124 17:14:15.876799 4777 generic.go:334] "Generic (PLEG): container finished" podID="0952d762-98cc-4c67-959a-772d127f6eb2" containerID="cbc3d8550cf41d4d9b2ce2194da843432ab6750e806ae46bf889432b690e0d57" exitCode=0 Nov 24 17:14:15 crc kubenswrapper[4777]: I1124 17:14:15.876862 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4brxh" event={"ID":"0952d762-98cc-4c67-959a-772d127f6eb2","Type":"ContainerDied","Data":"cbc3d8550cf41d4d9b2ce2194da843432ab6750e806ae46bf889432b690e0d57"} Nov 24 17:14:15 crc kubenswrapper[4777]: I1124 17:14:15.878592 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4brxh" event={"ID":"0952d762-98cc-4c67-959a-772d127f6eb2","Type":"ContainerStarted","Data":"1f7e2d81cd61dc686be30102aa9fc8e63b23d54ffcb29b018dec77c75c5e10d0"} Nov 24 17:14:17 crc kubenswrapper[4777]: I1124 17:14:17.809943 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-jxbj9" podUID="10516d2d-106f-4155-ab2c-c0a302ca5cbf" containerName="console" containerID="cri-o://102ae36ed9ce8ddf96acc06d1168c28ad8ba207eb6d34f0b39e8483ebc33be32" gracePeriod=15 Nov 24 17:14:17 crc kubenswrapper[4777]: I1124 17:14:17.937083 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:17 crc kubenswrapper[4777]: I1124 17:14:17.937565 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:19 crc kubenswrapper[4777]: I1124 17:14:19.002161 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9b95k" podUID="4670321e-4241-4686-8eaa-860bac08a44f" containerName="registry-server" probeResult="failure" output=< Nov 24 17:14:19 crc kubenswrapper[4777]: timeout: failed to connect service ":50051" within 1s Nov 24 17:14:19 crc kubenswrapper[4777]: > Nov 24 17:14:19 crc kubenswrapper[4777]: I1124 17:14:19.040608 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4"] Nov 24 17:14:19 crc kubenswrapper[4777]: I1124 17:14:19.042032 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" Nov 24 17:14:19 crc kubenswrapper[4777]: I1124 17:14:19.047175 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 17:14:19 crc kubenswrapper[4777]: I1124 17:14:19.055619 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/539a904e-3754-42f1-95ec-13140964520d-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4\" (UID: \"539a904e-3754-42f1-95ec-13140964520d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" Nov 24 17:14:19 crc kubenswrapper[4777]: I1124 17:14:19.055679 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/539a904e-3754-42f1-95ec-13140964520d-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4\" (UID: \"539a904e-3754-42f1-95ec-13140964520d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" Nov 24 17:14:19 crc kubenswrapper[4777]: I1124 17:14:19.055721 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl7nh\" (UniqueName: \"kubernetes.io/projected/539a904e-3754-42f1-95ec-13140964520d-kube-api-access-sl7nh\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4\" (UID: \"539a904e-3754-42f1-95ec-13140964520d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" Nov 24 17:14:19 crc kubenswrapper[4777]: I1124 17:14:19.059800 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4"] Nov 24 17:14:19 crc kubenswrapper[4777]: I1124 17:14:19.157237 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/539a904e-3754-42f1-95ec-13140964520d-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4\" (UID: \"539a904e-3754-42f1-95ec-13140964520d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" Nov 24 17:14:19 crc kubenswrapper[4777]: I1124 17:14:19.157318 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl7nh\" (UniqueName: \"kubernetes.io/projected/539a904e-3754-42f1-95ec-13140964520d-kube-api-access-sl7nh\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4\" (UID: \"539a904e-3754-42f1-95ec-13140964520d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" Nov 24 17:14:19 crc kubenswrapper[4777]: I1124 17:14:19.157372 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/539a904e-3754-42f1-95ec-13140964520d-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4\" (UID: \"539a904e-3754-42f1-95ec-13140964520d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" Nov 24 17:14:19 crc kubenswrapper[4777]: I1124 17:14:19.157820 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/539a904e-3754-42f1-95ec-13140964520d-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4\" (UID: \"539a904e-3754-42f1-95ec-13140964520d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" Nov 24 17:14:19 crc kubenswrapper[4777]: I1124 17:14:19.157831 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/539a904e-3754-42f1-95ec-13140964520d-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4\" (UID: \"539a904e-3754-42f1-95ec-13140964520d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" Nov 24 17:14:19 crc kubenswrapper[4777]: I1124 17:14:19.178473 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl7nh\" (UniqueName: \"kubernetes.io/projected/539a904e-3754-42f1-95ec-13140964520d-kube-api-access-sl7nh\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4\" (UID: \"539a904e-3754-42f1-95ec-13140964520d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" Nov 24 17:14:19 crc kubenswrapper[4777]: I1124 17:14:19.360476 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" Nov 24 17:14:20 crc kubenswrapper[4777]: I1124 17:14:20.289209 4777 patch_prober.go:28] interesting pod/console-f9d7485db-jxbj9 container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 24 17:14:20 crc kubenswrapper[4777]: I1124 17:14:20.289641 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-f9d7485db-jxbj9" podUID="10516d2d-106f-4155-ab2c-c0a302ca5cbf" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 24 17:14:21 crc kubenswrapper[4777]: I1124 17:14:21.076413 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4"] Nov 24 17:14:21 crc kubenswrapper[4777]: W1124 17:14:21.086541 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod539a904e_3754_42f1_95ec_13140964520d.slice/crio-68ff1f679f5b91d20de1e473c8421b6c9cd6aaa7bf80a9a3fed9270f062a193d WatchSource:0}: Error finding container 68ff1f679f5b91d20de1e473c8421b6c9cd6aaa7bf80a9a3fed9270f062a193d: Status 404 returned error can't find the container with id 68ff1f679f5b91d20de1e473c8421b6c9cd6aaa7bf80a9a3fed9270f062a193d Nov 24 17:14:21 crc kubenswrapper[4777]: I1124 17:14:21.929823 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-jxbj9_10516d2d-106f-4155-ab2c-c0a302ca5cbf/console/0.log" Nov 24 17:14:21 crc kubenswrapper[4777]: I1124 17:14:21.930289 4777 generic.go:334] "Generic (PLEG): container finished" podID="10516d2d-106f-4155-ab2c-c0a302ca5cbf" containerID="102ae36ed9ce8ddf96acc06d1168c28ad8ba207eb6d34f0b39e8483ebc33be32" exitCode=2 Nov 24 17:14:21 crc kubenswrapper[4777]: I1124 17:14:21.930389 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-jxbj9" event={"ID":"10516d2d-106f-4155-ab2c-c0a302ca5cbf","Type":"ContainerDied","Data":"102ae36ed9ce8ddf96acc06d1168c28ad8ba207eb6d34f0b39e8483ebc33be32"} Nov 24 17:14:21 crc kubenswrapper[4777]: I1124 17:14:21.930426 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-jxbj9" event={"ID":"10516d2d-106f-4155-ab2c-c0a302ca5cbf","Type":"ContainerDied","Data":"01960da356fb01e61c8a30a8cf978e1ce71ff26983c4ad232650a6b79a93c3e1"} Nov 24 17:14:21 crc kubenswrapper[4777]: I1124 17:14:21.930444 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01960da356fb01e61c8a30a8cf978e1ce71ff26983c4ad232650a6b79a93c3e1" Nov 24 17:14:21 crc kubenswrapper[4777]: I1124 17:14:21.958774 4777 generic.go:334] "Generic (PLEG): container finished" podID="0952d762-98cc-4c67-959a-772d127f6eb2" containerID="d42fa5fe1dee58e9120acbf73d8375a11ebed4c6dd01405821d7c29c45d90e2b" exitCode=0 Nov 24 17:14:21 crc kubenswrapper[4777]: I1124 17:14:21.958832 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4brxh" event={"ID":"0952d762-98cc-4c67-959a-772d127f6eb2","Type":"ContainerDied","Data":"d42fa5fe1dee58e9120acbf73d8375a11ebed4c6dd01405821d7c29c45d90e2b"} Nov 24 17:14:21 crc kubenswrapper[4777]: I1124 17:14:21.961667 4777 generic.go:334] "Generic (PLEG): container finished" podID="539a904e-3754-42f1-95ec-13140964520d" containerID="00ff50d9b7bfc073769cf833613733c53accde98302e553687542cad8dc96a58" exitCode=0 Nov 24 17:14:21 crc kubenswrapper[4777]: I1124 17:14:21.961730 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" event={"ID":"539a904e-3754-42f1-95ec-13140964520d","Type":"ContainerDied","Data":"00ff50d9b7bfc073769cf833613733c53accde98302e553687542cad8dc96a58"} Nov 24 17:14:21 crc kubenswrapper[4777]: I1124 17:14:21.961770 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" event={"ID":"539a904e-3754-42f1-95ec-13140964520d","Type":"ContainerStarted","Data":"68ff1f679f5b91d20de1e473c8421b6c9cd6aaa7bf80a9a3fed9270f062a193d"} Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.034393 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-jxbj9_10516d2d-106f-4155-ab2c-c0a302ca5cbf/console/0.log" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.034480 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.101841 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-config\") pod \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.101890 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-service-ca\") pod \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.101917 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-752sw\" (UniqueName: \"kubernetes.io/projected/10516d2d-106f-4155-ab2c-c0a302ca5cbf-kube-api-access-752sw\") pod \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.101946 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-oauth-config\") pod \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.102008 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-trusted-ca-bundle\") pod \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.102085 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-serving-cert\") pod \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.102136 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-oauth-serving-cert\") pod \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\" (UID: \"10516d2d-106f-4155-ab2c-c0a302ca5cbf\") " Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.102566 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-config" (OuterVolumeSpecName: "console-config") pod "10516d2d-106f-4155-ab2c-c0a302ca5cbf" (UID: "10516d2d-106f-4155-ab2c-c0a302ca5cbf"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.102740 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "10516d2d-106f-4155-ab2c-c0a302ca5cbf" (UID: "10516d2d-106f-4155-ab2c-c0a302ca5cbf"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.103018 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "10516d2d-106f-4155-ab2c-c0a302ca5cbf" (UID: "10516d2d-106f-4155-ab2c-c0a302ca5cbf"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.107750 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-service-ca" (OuterVolumeSpecName: "service-ca") pod "10516d2d-106f-4155-ab2c-c0a302ca5cbf" (UID: "10516d2d-106f-4155-ab2c-c0a302ca5cbf"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.114098 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "10516d2d-106f-4155-ab2c-c0a302ca5cbf" (UID: "10516d2d-106f-4155-ab2c-c0a302ca5cbf"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.119900 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10516d2d-106f-4155-ab2c-c0a302ca5cbf-kube-api-access-752sw" (OuterVolumeSpecName: "kube-api-access-752sw") pod "10516d2d-106f-4155-ab2c-c0a302ca5cbf" (UID: "10516d2d-106f-4155-ab2c-c0a302ca5cbf"). InnerVolumeSpecName "kube-api-access-752sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.130953 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "10516d2d-106f-4155-ab2c-c0a302ca5cbf" (UID: "10516d2d-106f-4155-ab2c-c0a302ca5cbf"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.204543 4777 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.204589 4777 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.204600 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-752sw\" (UniqueName: \"kubernetes.io/projected/10516d2d-106f-4155-ab2c-c0a302ca5cbf-kube-api-access-752sw\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.204614 4777 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.204625 4777 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.204634 4777 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/10516d2d-106f-4155-ab2c-c0a302ca5cbf-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.204644 4777 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/10516d2d-106f-4155-ab2c-c0a302ca5cbf-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.978019 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-jxbj9" Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.978002 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4brxh" event={"ID":"0952d762-98cc-4c67-959a-772d127f6eb2","Type":"ContainerStarted","Data":"831799c97a1f8356ee4d0ae8b8b5728cdc507c0eef3f10e24a6a1cd23b30eb55"} Nov 24 17:14:22 crc kubenswrapper[4777]: I1124 17:14:22.999579 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4brxh" podStartSLOduration=2.503163301 podStartE2EDuration="8.999559283s" podCreationTimestamp="2025-11-24 17:14:14 +0000 UTC" firstStartedPulling="2025-11-24 17:14:15.878933011 +0000 UTC m=+814.038128060" lastFinishedPulling="2025-11-24 17:14:22.375328993 +0000 UTC m=+820.534524042" observedRunningTime="2025-11-24 17:14:22.999341687 +0000 UTC m=+821.158536766" watchObservedRunningTime="2025-11-24 17:14:22.999559283 +0000 UTC m=+821.158754352" Nov 24 17:14:23 crc kubenswrapper[4777]: I1124 17:14:23.017450 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-jxbj9"] Nov 24 17:14:23 crc kubenswrapper[4777]: I1124 17:14:23.021043 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-jxbj9"] Nov 24 17:14:23 crc kubenswrapper[4777]: I1124 17:14:23.255235 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10516d2d-106f-4155-ab2c-c0a302ca5cbf" path="/var/lib/kubelet/pods/10516d2d-106f-4155-ab2c-c0a302ca5cbf/volumes" Nov 24 17:14:23 crc kubenswrapper[4777]: I1124 17:14:23.987247 4777 generic.go:334] "Generic (PLEG): container finished" podID="539a904e-3754-42f1-95ec-13140964520d" containerID="e55b72fe49a29c39b4fd2c0c334b9e02ef9dd63d6fbf65623f2454863adf7355" exitCode=0 Nov 24 17:14:23 crc kubenswrapper[4777]: I1124 17:14:23.987369 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" event={"ID":"539a904e-3754-42f1-95ec-13140964520d","Type":"ContainerDied","Data":"e55b72fe49a29c39b4fd2c0c334b9e02ef9dd63d6fbf65623f2454863adf7355"} Nov 24 17:14:24 crc kubenswrapper[4777]: I1124 17:14:24.534269 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:24 crc kubenswrapper[4777]: I1124 17:14:24.534516 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:24 crc kubenswrapper[4777]: I1124 17:14:24.606258 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:24 crc kubenswrapper[4777]: I1124 17:14:24.996805 4777 generic.go:334] "Generic (PLEG): container finished" podID="539a904e-3754-42f1-95ec-13140964520d" containerID="6a87a884eea607c89ce27e75fd859bc5ce801b96d005c360da0b9de0a20ec1e6" exitCode=0 Nov 24 17:14:24 crc kubenswrapper[4777]: I1124 17:14:24.996846 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" event={"ID":"539a904e-3754-42f1-95ec-13140964520d","Type":"ContainerDied","Data":"6a87a884eea607c89ce27e75fd859bc5ce801b96d005c360da0b9de0a20ec1e6"} Nov 24 17:14:26 crc kubenswrapper[4777]: I1124 17:14:26.311744 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" Nov 24 17:14:26 crc kubenswrapper[4777]: I1124 17:14:26.417749 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/539a904e-3754-42f1-95ec-13140964520d-bundle\") pod \"539a904e-3754-42f1-95ec-13140964520d\" (UID: \"539a904e-3754-42f1-95ec-13140964520d\") " Nov 24 17:14:26 crc kubenswrapper[4777]: I1124 17:14:26.417848 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sl7nh\" (UniqueName: \"kubernetes.io/projected/539a904e-3754-42f1-95ec-13140964520d-kube-api-access-sl7nh\") pod \"539a904e-3754-42f1-95ec-13140964520d\" (UID: \"539a904e-3754-42f1-95ec-13140964520d\") " Nov 24 17:14:26 crc kubenswrapper[4777]: I1124 17:14:26.418002 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/539a904e-3754-42f1-95ec-13140964520d-util\") pod \"539a904e-3754-42f1-95ec-13140964520d\" (UID: \"539a904e-3754-42f1-95ec-13140964520d\") " Nov 24 17:14:26 crc kubenswrapper[4777]: I1124 17:14:26.419250 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/539a904e-3754-42f1-95ec-13140964520d-bundle" (OuterVolumeSpecName: "bundle") pod "539a904e-3754-42f1-95ec-13140964520d" (UID: "539a904e-3754-42f1-95ec-13140964520d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:14:26 crc kubenswrapper[4777]: I1124 17:14:26.423722 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/539a904e-3754-42f1-95ec-13140964520d-kube-api-access-sl7nh" (OuterVolumeSpecName: "kube-api-access-sl7nh") pod "539a904e-3754-42f1-95ec-13140964520d" (UID: "539a904e-3754-42f1-95ec-13140964520d"). InnerVolumeSpecName "kube-api-access-sl7nh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:14:26 crc kubenswrapper[4777]: I1124 17:14:26.439341 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/539a904e-3754-42f1-95ec-13140964520d-util" (OuterVolumeSpecName: "util") pod "539a904e-3754-42f1-95ec-13140964520d" (UID: "539a904e-3754-42f1-95ec-13140964520d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:14:26 crc kubenswrapper[4777]: I1124 17:14:26.520104 4777 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/539a904e-3754-42f1-95ec-13140964520d-util\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:26 crc kubenswrapper[4777]: I1124 17:14:26.520435 4777 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/539a904e-3754-42f1-95ec-13140964520d-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:26 crc kubenswrapper[4777]: I1124 17:14:26.520449 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sl7nh\" (UniqueName: \"kubernetes.io/projected/539a904e-3754-42f1-95ec-13140964520d-kube-api-access-sl7nh\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:27 crc kubenswrapper[4777]: I1124 17:14:27.023802 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" event={"ID":"539a904e-3754-42f1-95ec-13140964520d","Type":"ContainerDied","Data":"68ff1f679f5b91d20de1e473c8421b6c9cd6aaa7bf80a9a3fed9270f062a193d"} Nov 24 17:14:27 crc kubenswrapper[4777]: I1124 17:14:27.023859 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68ff1f679f5b91d20de1e473c8421b6c9cd6aaa7bf80a9a3fed9270f062a193d" Nov 24 17:14:27 crc kubenswrapper[4777]: I1124 17:14:27.023952 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4" Nov 24 17:14:27 crc kubenswrapper[4777]: I1124 17:14:27.999847 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:28 crc kubenswrapper[4777]: I1124 17:14:28.049316 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:28 crc kubenswrapper[4777]: I1124 17:14:28.981793 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9b95k"] Nov 24 17:14:29 crc kubenswrapper[4777]: I1124 17:14:29.038790 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9b95k" podUID="4670321e-4241-4686-8eaa-860bac08a44f" containerName="registry-server" containerID="cri-o://0a51dbe85c1afef6b369cac0aaab762498eb34832bce9b55d06a0f5b8757732c" gracePeriod=2 Nov 24 17:14:29 crc kubenswrapper[4777]: I1124 17:14:29.498901 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:29 crc kubenswrapper[4777]: I1124 17:14:29.665611 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmpng\" (UniqueName: \"kubernetes.io/projected/4670321e-4241-4686-8eaa-860bac08a44f-kube-api-access-vmpng\") pod \"4670321e-4241-4686-8eaa-860bac08a44f\" (UID: \"4670321e-4241-4686-8eaa-860bac08a44f\") " Nov 24 17:14:29 crc kubenswrapper[4777]: I1124 17:14:29.665683 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4670321e-4241-4686-8eaa-860bac08a44f-catalog-content\") pod \"4670321e-4241-4686-8eaa-860bac08a44f\" (UID: \"4670321e-4241-4686-8eaa-860bac08a44f\") " Nov 24 17:14:29 crc kubenswrapper[4777]: I1124 17:14:29.665783 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4670321e-4241-4686-8eaa-860bac08a44f-utilities\") pod \"4670321e-4241-4686-8eaa-860bac08a44f\" (UID: \"4670321e-4241-4686-8eaa-860bac08a44f\") " Nov 24 17:14:29 crc kubenswrapper[4777]: I1124 17:14:29.666510 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4670321e-4241-4686-8eaa-860bac08a44f-utilities" (OuterVolumeSpecName: "utilities") pod "4670321e-4241-4686-8eaa-860bac08a44f" (UID: "4670321e-4241-4686-8eaa-860bac08a44f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:14:29 crc kubenswrapper[4777]: I1124 17:14:29.682178 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4670321e-4241-4686-8eaa-860bac08a44f-kube-api-access-vmpng" (OuterVolumeSpecName: "kube-api-access-vmpng") pod "4670321e-4241-4686-8eaa-860bac08a44f" (UID: "4670321e-4241-4686-8eaa-860bac08a44f"). InnerVolumeSpecName "kube-api-access-vmpng". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:14:29 crc kubenswrapper[4777]: I1124 17:14:29.767662 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4670321e-4241-4686-8eaa-860bac08a44f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:29 crc kubenswrapper[4777]: I1124 17:14:29.767698 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmpng\" (UniqueName: \"kubernetes.io/projected/4670321e-4241-4686-8eaa-860bac08a44f-kube-api-access-vmpng\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:29 crc kubenswrapper[4777]: I1124 17:14:29.785791 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4670321e-4241-4686-8eaa-860bac08a44f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4670321e-4241-4686-8eaa-860bac08a44f" (UID: "4670321e-4241-4686-8eaa-860bac08a44f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:14:29 crc kubenswrapper[4777]: I1124 17:14:29.869354 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4670321e-4241-4686-8eaa-860bac08a44f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:30 crc kubenswrapper[4777]: I1124 17:14:30.045985 4777 generic.go:334] "Generic (PLEG): container finished" podID="4670321e-4241-4686-8eaa-860bac08a44f" containerID="0a51dbe85c1afef6b369cac0aaab762498eb34832bce9b55d06a0f5b8757732c" exitCode=0 Nov 24 17:14:30 crc kubenswrapper[4777]: I1124 17:14:30.046061 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9b95k" Nov 24 17:14:30 crc kubenswrapper[4777]: I1124 17:14:30.046093 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9b95k" event={"ID":"4670321e-4241-4686-8eaa-860bac08a44f","Type":"ContainerDied","Data":"0a51dbe85c1afef6b369cac0aaab762498eb34832bce9b55d06a0f5b8757732c"} Nov 24 17:14:30 crc kubenswrapper[4777]: I1124 17:14:30.046393 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9b95k" event={"ID":"4670321e-4241-4686-8eaa-860bac08a44f","Type":"ContainerDied","Data":"aeb3796be539fcbbebb4163fc6b7afe50ec6145e2800de1919d39591986947b2"} Nov 24 17:14:30 crc kubenswrapper[4777]: I1124 17:14:30.046419 4777 scope.go:117] "RemoveContainer" containerID="0a51dbe85c1afef6b369cac0aaab762498eb34832bce9b55d06a0f5b8757732c" Nov 24 17:14:30 crc kubenswrapper[4777]: I1124 17:14:30.066439 4777 scope.go:117] "RemoveContainer" containerID="dbd3b277144c52f170d546ba6ae621228c8b02c6a3aea9599c431a16f7573073" Nov 24 17:14:30 crc kubenswrapper[4777]: I1124 17:14:30.097041 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9b95k"] Nov 24 17:14:30 crc kubenswrapper[4777]: I1124 17:14:30.097199 4777 scope.go:117] "RemoveContainer" containerID="1b2fc694064baa7c4c4173e3df3122a28a3a224863e55d7246d4bff6f9afe0c9" Nov 24 17:14:30 crc kubenswrapper[4777]: I1124 17:14:30.113380 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9b95k"] Nov 24 17:14:30 crc kubenswrapper[4777]: I1124 17:14:30.132630 4777 scope.go:117] "RemoveContainer" containerID="0a51dbe85c1afef6b369cac0aaab762498eb34832bce9b55d06a0f5b8757732c" Nov 24 17:14:30 crc kubenswrapper[4777]: E1124 17:14:30.134097 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a51dbe85c1afef6b369cac0aaab762498eb34832bce9b55d06a0f5b8757732c\": container with ID starting with 0a51dbe85c1afef6b369cac0aaab762498eb34832bce9b55d06a0f5b8757732c not found: ID does not exist" containerID="0a51dbe85c1afef6b369cac0aaab762498eb34832bce9b55d06a0f5b8757732c" Nov 24 17:14:30 crc kubenswrapper[4777]: I1124 17:14:30.134128 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a51dbe85c1afef6b369cac0aaab762498eb34832bce9b55d06a0f5b8757732c"} err="failed to get container status \"0a51dbe85c1afef6b369cac0aaab762498eb34832bce9b55d06a0f5b8757732c\": rpc error: code = NotFound desc = could not find container \"0a51dbe85c1afef6b369cac0aaab762498eb34832bce9b55d06a0f5b8757732c\": container with ID starting with 0a51dbe85c1afef6b369cac0aaab762498eb34832bce9b55d06a0f5b8757732c not found: ID does not exist" Nov 24 17:14:30 crc kubenswrapper[4777]: I1124 17:14:30.134150 4777 scope.go:117] "RemoveContainer" containerID="dbd3b277144c52f170d546ba6ae621228c8b02c6a3aea9599c431a16f7573073" Nov 24 17:14:30 crc kubenswrapper[4777]: E1124 17:14:30.134435 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbd3b277144c52f170d546ba6ae621228c8b02c6a3aea9599c431a16f7573073\": container with ID starting with dbd3b277144c52f170d546ba6ae621228c8b02c6a3aea9599c431a16f7573073 not found: ID does not exist" containerID="dbd3b277144c52f170d546ba6ae621228c8b02c6a3aea9599c431a16f7573073" Nov 24 17:14:30 crc kubenswrapper[4777]: I1124 17:14:30.134481 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbd3b277144c52f170d546ba6ae621228c8b02c6a3aea9599c431a16f7573073"} err="failed to get container status \"dbd3b277144c52f170d546ba6ae621228c8b02c6a3aea9599c431a16f7573073\": rpc error: code = NotFound desc = could not find container \"dbd3b277144c52f170d546ba6ae621228c8b02c6a3aea9599c431a16f7573073\": container with ID starting with dbd3b277144c52f170d546ba6ae621228c8b02c6a3aea9599c431a16f7573073 not found: ID does not exist" Nov 24 17:14:30 crc kubenswrapper[4777]: I1124 17:14:30.134510 4777 scope.go:117] "RemoveContainer" containerID="1b2fc694064baa7c4c4173e3df3122a28a3a224863e55d7246d4bff6f9afe0c9" Nov 24 17:14:30 crc kubenswrapper[4777]: E1124 17:14:30.134740 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b2fc694064baa7c4c4173e3df3122a28a3a224863e55d7246d4bff6f9afe0c9\": container with ID starting with 1b2fc694064baa7c4c4173e3df3122a28a3a224863e55d7246d4bff6f9afe0c9 not found: ID does not exist" containerID="1b2fc694064baa7c4c4173e3df3122a28a3a224863e55d7246d4bff6f9afe0c9" Nov 24 17:14:30 crc kubenswrapper[4777]: I1124 17:14:30.134761 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b2fc694064baa7c4c4173e3df3122a28a3a224863e55d7246d4bff6f9afe0c9"} err="failed to get container status \"1b2fc694064baa7c4c4173e3df3122a28a3a224863e55d7246d4bff6f9afe0c9\": rpc error: code = NotFound desc = could not find container \"1b2fc694064baa7c4c4173e3df3122a28a3a224863e55d7246d4bff6f9afe0c9\": container with ID starting with 1b2fc694064baa7c4c4173e3df3122a28a3a224863e55d7246d4bff6f9afe0c9 not found: ID does not exist" Nov 24 17:14:31 crc kubenswrapper[4777]: I1124 17:14:31.254793 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4670321e-4241-4686-8eaa-860bac08a44f" path="/var/lib/kubelet/pods/4670321e-4241-4686-8eaa-860bac08a44f/volumes" Nov 24 17:14:34 crc kubenswrapper[4777]: I1124 17:14:34.578765 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.089706 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb"] Nov 24 17:14:36 crc kubenswrapper[4777]: E1124 17:14:36.090016 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4670321e-4241-4686-8eaa-860bac08a44f" containerName="extract-utilities" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.090033 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="4670321e-4241-4686-8eaa-860bac08a44f" containerName="extract-utilities" Nov 24 17:14:36 crc kubenswrapper[4777]: E1124 17:14:36.090046 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4670321e-4241-4686-8eaa-860bac08a44f" containerName="registry-server" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.090054 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="4670321e-4241-4686-8eaa-860bac08a44f" containerName="registry-server" Nov 24 17:14:36 crc kubenswrapper[4777]: E1124 17:14:36.090072 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="539a904e-3754-42f1-95ec-13140964520d" containerName="pull" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.090082 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="539a904e-3754-42f1-95ec-13140964520d" containerName="pull" Nov 24 17:14:36 crc kubenswrapper[4777]: E1124 17:14:36.090100 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="539a904e-3754-42f1-95ec-13140964520d" containerName="extract" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.090107 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="539a904e-3754-42f1-95ec-13140964520d" containerName="extract" Nov 24 17:14:36 crc kubenswrapper[4777]: E1124 17:14:36.090118 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4670321e-4241-4686-8eaa-860bac08a44f" containerName="extract-content" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.090126 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="4670321e-4241-4686-8eaa-860bac08a44f" containerName="extract-content" Nov 24 17:14:36 crc kubenswrapper[4777]: E1124 17:14:36.090136 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="539a904e-3754-42f1-95ec-13140964520d" containerName="util" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.090143 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="539a904e-3754-42f1-95ec-13140964520d" containerName="util" Nov 24 17:14:36 crc kubenswrapper[4777]: E1124 17:14:36.090153 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10516d2d-106f-4155-ab2c-c0a302ca5cbf" containerName="console" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.090160 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="10516d2d-106f-4155-ab2c-c0a302ca5cbf" containerName="console" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.090281 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="4670321e-4241-4686-8eaa-860bac08a44f" containerName="registry-server" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.090305 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="10516d2d-106f-4155-ab2c-c0a302ca5cbf" containerName="console" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.090314 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="539a904e-3754-42f1-95ec-13140964520d" containerName="extract" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.091094 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.093620 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.093665 4777 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.093781 4777 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-nv98d" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.093871 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.093929 4777 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.110883 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb"] Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.269339 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/364328a8-4ddb-4c3c-96bd-3a87ef271678-webhook-cert\") pod \"metallb-operator-controller-manager-7444c76584-h5nzb\" (UID: \"364328a8-4ddb-4c3c-96bd-3a87ef271678\") " pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.269425 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/364328a8-4ddb-4c3c-96bd-3a87ef271678-apiservice-cert\") pod \"metallb-operator-controller-manager-7444c76584-h5nzb\" (UID: \"364328a8-4ddb-4c3c-96bd-3a87ef271678\") " pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.269641 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dgbf\" (UniqueName: \"kubernetes.io/projected/364328a8-4ddb-4c3c-96bd-3a87ef271678-kube-api-access-2dgbf\") pod \"metallb-operator-controller-manager-7444c76584-h5nzb\" (UID: \"364328a8-4ddb-4c3c-96bd-3a87ef271678\") " pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.338602 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd"] Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.339283 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.343400 4777 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.343456 4777 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.343947 4777 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-zjbn7" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.354748 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd"] Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.370469 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dgbf\" (UniqueName: \"kubernetes.io/projected/364328a8-4ddb-4c3c-96bd-3a87ef271678-kube-api-access-2dgbf\") pod \"metallb-operator-controller-manager-7444c76584-h5nzb\" (UID: \"364328a8-4ddb-4c3c-96bd-3a87ef271678\") " pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.370539 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/364328a8-4ddb-4c3c-96bd-3a87ef271678-webhook-cert\") pod \"metallb-operator-controller-manager-7444c76584-h5nzb\" (UID: \"364328a8-4ddb-4c3c-96bd-3a87ef271678\") " pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.370603 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/364328a8-4ddb-4c3c-96bd-3a87ef271678-apiservice-cert\") pod \"metallb-operator-controller-manager-7444c76584-h5nzb\" (UID: \"364328a8-4ddb-4c3c-96bd-3a87ef271678\") " pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.392745 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/364328a8-4ddb-4c3c-96bd-3a87ef271678-webhook-cert\") pod \"metallb-operator-controller-manager-7444c76584-h5nzb\" (UID: \"364328a8-4ddb-4c3c-96bd-3a87ef271678\") " pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.400985 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/364328a8-4ddb-4c3c-96bd-3a87ef271678-apiservice-cert\") pod \"metallb-operator-controller-manager-7444c76584-h5nzb\" (UID: \"364328a8-4ddb-4c3c-96bd-3a87ef271678\") " pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.404123 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dgbf\" (UniqueName: \"kubernetes.io/projected/364328a8-4ddb-4c3c-96bd-3a87ef271678-kube-api-access-2dgbf\") pod \"metallb-operator-controller-manager-7444c76584-h5nzb\" (UID: \"364328a8-4ddb-4c3c-96bd-3a87ef271678\") " pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.420297 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.472013 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/61277f6c-5e55-4fdb-a482-e9b3a91edd7f-apiservice-cert\") pod \"metallb-operator-webhook-server-7578594d7b-kn9dd\" (UID: \"61277f6c-5e55-4fdb-a482-e9b3a91edd7f\") " pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.472086 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6ws6\" (UniqueName: \"kubernetes.io/projected/61277f6c-5e55-4fdb-a482-e9b3a91edd7f-kube-api-access-v6ws6\") pod \"metallb-operator-webhook-server-7578594d7b-kn9dd\" (UID: \"61277f6c-5e55-4fdb-a482-e9b3a91edd7f\") " pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.472111 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/61277f6c-5e55-4fdb-a482-e9b3a91edd7f-webhook-cert\") pod \"metallb-operator-webhook-server-7578594d7b-kn9dd\" (UID: \"61277f6c-5e55-4fdb-a482-e9b3a91edd7f\") " pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.574849 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/61277f6c-5e55-4fdb-a482-e9b3a91edd7f-apiservice-cert\") pod \"metallb-operator-webhook-server-7578594d7b-kn9dd\" (UID: \"61277f6c-5e55-4fdb-a482-e9b3a91edd7f\") " pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.574931 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6ws6\" (UniqueName: \"kubernetes.io/projected/61277f6c-5e55-4fdb-a482-e9b3a91edd7f-kube-api-access-v6ws6\") pod \"metallb-operator-webhook-server-7578594d7b-kn9dd\" (UID: \"61277f6c-5e55-4fdb-a482-e9b3a91edd7f\") " pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.574964 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/61277f6c-5e55-4fdb-a482-e9b3a91edd7f-webhook-cert\") pod \"metallb-operator-webhook-server-7578594d7b-kn9dd\" (UID: \"61277f6c-5e55-4fdb-a482-e9b3a91edd7f\") " pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.582882 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/61277f6c-5e55-4fdb-a482-e9b3a91edd7f-webhook-cert\") pod \"metallb-operator-webhook-server-7578594d7b-kn9dd\" (UID: \"61277f6c-5e55-4fdb-a482-e9b3a91edd7f\") " pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.585353 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/61277f6c-5e55-4fdb-a482-e9b3a91edd7f-apiservice-cert\") pod \"metallb-operator-webhook-server-7578594d7b-kn9dd\" (UID: \"61277f6c-5e55-4fdb-a482-e9b3a91edd7f\") " pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.604268 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6ws6\" (UniqueName: \"kubernetes.io/projected/61277f6c-5e55-4fdb-a482-e9b3a91edd7f-kube-api-access-v6ws6\") pod \"metallb-operator-webhook-server-7578594d7b-kn9dd\" (UID: \"61277f6c-5e55-4fdb-a482-e9b3a91edd7f\") " pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.652636 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.811390 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lt7fn"] Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.813150 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.826639 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lt7fn"] Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.879264 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb"] Nov 24 17:14:36 crc kubenswrapper[4777]: W1124 17:14:36.901198 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod364328a8_4ddb_4c3c_96bd_3a87ef271678.slice/crio-8b2e0215d52d904f894fe047b5d0a8d72221b5a8ed5cc78b1543405c0f088380 WatchSource:0}: Error finding container 8b2e0215d52d904f894fe047b5d0a8d72221b5a8ed5cc78b1543405c0f088380: Status 404 returned error can't find the container with id 8b2e0215d52d904f894fe047b5d0a8d72221b5a8ed5cc78b1543405c0f088380 Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.982895 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75460ef2-43a2-4488-85eb-6f58ab746e0a-utilities\") pod \"redhat-marketplace-lt7fn\" (UID: \"75460ef2-43a2-4488-85eb-6f58ab746e0a\") " pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.982957 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75460ef2-43a2-4488-85eb-6f58ab746e0a-catalog-content\") pod \"redhat-marketplace-lt7fn\" (UID: \"75460ef2-43a2-4488-85eb-6f58ab746e0a\") " pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.983021 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7mwd\" (UniqueName: \"kubernetes.io/projected/75460ef2-43a2-4488-85eb-6f58ab746e0a-kube-api-access-m7mwd\") pod \"redhat-marketplace-lt7fn\" (UID: \"75460ef2-43a2-4488-85eb-6f58ab746e0a\") " pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:36 crc kubenswrapper[4777]: I1124 17:14:36.987216 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd"] Nov 24 17:14:36 crc kubenswrapper[4777]: W1124 17:14:36.997117 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61277f6c_5e55_4fdb_a482_e9b3a91edd7f.slice/crio-27e26a2cf21986f3481388dc218824df97600badaf469bc41c3cba582ee92dad WatchSource:0}: Error finding container 27e26a2cf21986f3481388dc218824df97600badaf469bc41c3cba582ee92dad: Status 404 returned error can't find the container with id 27e26a2cf21986f3481388dc218824df97600badaf469bc41c3cba582ee92dad Nov 24 17:14:37 crc kubenswrapper[4777]: I1124 17:14:37.083786 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75460ef2-43a2-4488-85eb-6f58ab746e0a-utilities\") pod \"redhat-marketplace-lt7fn\" (UID: \"75460ef2-43a2-4488-85eb-6f58ab746e0a\") " pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:37 crc kubenswrapper[4777]: I1124 17:14:37.083844 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75460ef2-43a2-4488-85eb-6f58ab746e0a-catalog-content\") pod \"redhat-marketplace-lt7fn\" (UID: \"75460ef2-43a2-4488-85eb-6f58ab746e0a\") " pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:37 crc kubenswrapper[4777]: I1124 17:14:37.083888 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7mwd\" (UniqueName: \"kubernetes.io/projected/75460ef2-43a2-4488-85eb-6f58ab746e0a-kube-api-access-m7mwd\") pod \"redhat-marketplace-lt7fn\" (UID: \"75460ef2-43a2-4488-85eb-6f58ab746e0a\") " pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:37 crc kubenswrapper[4777]: I1124 17:14:37.084384 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75460ef2-43a2-4488-85eb-6f58ab746e0a-utilities\") pod \"redhat-marketplace-lt7fn\" (UID: \"75460ef2-43a2-4488-85eb-6f58ab746e0a\") " pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:37 crc kubenswrapper[4777]: I1124 17:14:37.084427 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75460ef2-43a2-4488-85eb-6f58ab746e0a-catalog-content\") pod \"redhat-marketplace-lt7fn\" (UID: \"75460ef2-43a2-4488-85eb-6f58ab746e0a\") " pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:37 crc kubenswrapper[4777]: I1124 17:14:37.094776 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" event={"ID":"364328a8-4ddb-4c3c-96bd-3a87ef271678","Type":"ContainerStarted","Data":"8b2e0215d52d904f894fe047b5d0a8d72221b5a8ed5cc78b1543405c0f088380"} Nov 24 17:14:37 crc kubenswrapper[4777]: I1124 17:14:37.095861 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" event={"ID":"61277f6c-5e55-4fdb-a482-e9b3a91edd7f","Type":"ContainerStarted","Data":"27e26a2cf21986f3481388dc218824df97600badaf469bc41c3cba582ee92dad"} Nov 24 17:14:37 crc kubenswrapper[4777]: I1124 17:14:37.104310 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7mwd\" (UniqueName: \"kubernetes.io/projected/75460ef2-43a2-4488-85eb-6f58ab746e0a-kube-api-access-m7mwd\") pod \"redhat-marketplace-lt7fn\" (UID: \"75460ef2-43a2-4488-85eb-6f58ab746e0a\") " pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:37 crc kubenswrapper[4777]: I1124 17:14:37.131295 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:37 crc kubenswrapper[4777]: I1124 17:14:37.547634 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lt7fn"] Nov 24 17:14:38 crc kubenswrapper[4777]: I1124 17:14:38.105120 4777 generic.go:334] "Generic (PLEG): container finished" podID="75460ef2-43a2-4488-85eb-6f58ab746e0a" containerID="d713eda52aeff3d82ca51ffd43cdd191f319ab56306bf175018a1617712627fb" exitCode=0 Nov 24 17:14:38 crc kubenswrapper[4777]: I1124 17:14:38.105172 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lt7fn" event={"ID":"75460ef2-43a2-4488-85eb-6f58ab746e0a","Type":"ContainerDied","Data":"d713eda52aeff3d82ca51ffd43cdd191f319ab56306bf175018a1617712627fb"} Nov 24 17:14:38 crc kubenswrapper[4777]: I1124 17:14:38.105201 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lt7fn" event={"ID":"75460ef2-43a2-4488-85eb-6f58ab746e0a","Type":"ContainerStarted","Data":"468498c1196b7eb45dbaf19b39e80d6f23bae6a9d48665d60f5325fa985b49fe"} Nov 24 17:14:40 crc kubenswrapper[4777]: I1124 17:14:40.127558 4777 generic.go:334] "Generic (PLEG): container finished" podID="75460ef2-43a2-4488-85eb-6f58ab746e0a" containerID="42cce577b6d8e8a3cf374c539beb4638e468583140d5974fa769f7e371c440f1" exitCode=0 Nov 24 17:14:40 crc kubenswrapper[4777]: I1124 17:14:40.127686 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lt7fn" event={"ID":"75460ef2-43a2-4488-85eb-6f58ab746e0a","Type":"ContainerDied","Data":"42cce577b6d8e8a3cf374c539beb4638e468583140d5974fa769f7e371c440f1"} Nov 24 17:14:41 crc kubenswrapper[4777]: I1124 17:14:41.388150 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4brxh"] Nov 24 17:14:41 crc kubenswrapper[4777]: I1124 17:14:41.388649 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4brxh" podUID="0952d762-98cc-4c67-959a-772d127f6eb2" containerName="registry-server" containerID="cri-o://831799c97a1f8356ee4d0ae8b8b5728cdc507c0eef3f10e24a6a1cd23b30eb55" gracePeriod=2 Nov 24 17:14:42 crc kubenswrapper[4777]: I1124 17:14:42.144633 4777 generic.go:334] "Generic (PLEG): container finished" podID="0952d762-98cc-4c67-959a-772d127f6eb2" containerID="831799c97a1f8356ee4d0ae8b8b5728cdc507c0eef3f10e24a6a1cd23b30eb55" exitCode=0 Nov 24 17:14:42 crc kubenswrapper[4777]: I1124 17:14:42.144727 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4brxh" event={"ID":"0952d762-98cc-4c67-959a-772d127f6eb2","Type":"ContainerDied","Data":"831799c97a1f8356ee4d0ae8b8b5728cdc507c0eef3f10e24a6a1cd23b30eb55"} Nov 24 17:14:42 crc kubenswrapper[4777]: I1124 17:14:42.874303 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:42 crc kubenswrapper[4777]: I1124 17:14:42.979741 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0952d762-98cc-4c67-959a-772d127f6eb2-utilities\") pod \"0952d762-98cc-4c67-959a-772d127f6eb2\" (UID: \"0952d762-98cc-4c67-959a-772d127f6eb2\") " Nov 24 17:14:42 crc kubenswrapper[4777]: I1124 17:14:42.979850 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0952d762-98cc-4c67-959a-772d127f6eb2-catalog-content\") pod \"0952d762-98cc-4c67-959a-772d127f6eb2\" (UID: \"0952d762-98cc-4c67-959a-772d127f6eb2\") " Nov 24 17:14:42 crc kubenswrapper[4777]: I1124 17:14:42.979897 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cj89\" (UniqueName: \"kubernetes.io/projected/0952d762-98cc-4c67-959a-772d127f6eb2-kube-api-access-5cj89\") pod \"0952d762-98cc-4c67-959a-772d127f6eb2\" (UID: \"0952d762-98cc-4c67-959a-772d127f6eb2\") " Nov 24 17:14:42 crc kubenswrapper[4777]: I1124 17:14:42.980606 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0952d762-98cc-4c67-959a-772d127f6eb2-utilities" (OuterVolumeSpecName: "utilities") pod "0952d762-98cc-4c67-959a-772d127f6eb2" (UID: "0952d762-98cc-4c67-959a-772d127f6eb2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:14:42 crc kubenswrapper[4777]: I1124 17:14:42.987079 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0952d762-98cc-4c67-959a-772d127f6eb2-kube-api-access-5cj89" (OuterVolumeSpecName: "kube-api-access-5cj89") pod "0952d762-98cc-4c67-959a-772d127f6eb2" (UID: "0952d762-98cc-4c67-959a-772d127f6eb2"). InnerVolumeSpecName "kube-api-access-5cj89". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.025447 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0952d762-98cc-4c67-959a-772d127f6eb2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0952d762-98cc-4c67-959a-772d127f6eb2" (UID: "0952d762-98cc-4c67-959a-772d127f6eb2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.082017 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0952d762-98cc-4c67-959a-772d127f6eb2-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.082054 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0952d762-98cc-4c67-959a-772d127f6eb2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.082069 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cj89\" (UniqueName: \"kubernetes.io/projected/0952d762-98cc-4c67-959a-772d127f6eb2-kube-api-access-5cj89\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.165161 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" event={"ID":"61277f6c-5e55-4fdb-a482-e9b3a91edd7f","Type":"ContainerStarted","Data":"16977b0f7ea1908c70d019b30e325b8dafcfdbb1fb49b6e25c6f9311498f4537"} Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.165251 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.167752 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4brxh" event={"ID":"0952d762-98cc-4c67-959a-772d127f6eb2","Type":"ContainerDied","Data":"1f7e2d81cd61dc686be30102aa9fc8e63b23d54ffcb29b018dec77c75c5e10d0"} Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.167793 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4brxh" Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.167808 4777 scope.go:117] "RemoveContainer" containerID="831799c97a1f8356ee4d0ae8b8b5728cdc507c0eef3f10e24a6a1cd23b30eb55" Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.172630 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" event={"ID":"364328a8-4ddb-4c3c-96bd-3a87ef271678","Type":"ContainerStarted","Data":"7ba91b09c92e8c2e896a1664fd12197522cb12482c3c6d48836cfd2c5fec1a67"} Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.172789 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.182092 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lt7fn" event={"ID":"75460ef2-43a2-4488-85eb-6f58ab746e0a","Type":"ContainerStarted","Data":"b440ba388414776d5107678ab7622d7f625ead43cd6c9868ee0d7624861d3caa"} Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.198489 4777 scope.go:117] "RemoveContainer" containerID="d42fa5fe1dee58e9120acbf73d8375a11ebed4c6dd01405821d7c29c45d90e2b" Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.218817 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lt7fn" podStartSLOduration=2.688695116 podStartE2EDuration="7.218801839s" podCreationTimestamp="2025-11-24 17:14:36 +0000 UTC" firstStartedPulling="2025-11-24 17:14:38.107288666 +0000 UTC m=+836.266483715" lastFinishedPulling="2025-11-24 17:14:42.637395379 +0000 UTC m=+840.796590438" observedRunningTime="2025-11-24 17:14:43.217172632 +0000 UTC m=+841.376367691" watchObservedRunningTime="2025-11-24 17:14:43.218801839 +0000 UTC m=+841.377996898" Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.219119 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" podStartSLOduration=1.55165825 podStartE2EDuration="7.219114559s" podCreationTimestamp="2025-11-24 17:14:36 +0000 UTC" firstStartedPulling="2025-11-24 17:14:36.998702302 +0000 UTC m=+835.157897341" lastFinishedPulling="2025-11-24 17:14:42.666158601 +0000 UTC m=+840.825353650" observedRunningTime="2025-11-24 17:14:43.193009123 +0000 UTC m=+841.352204182" watchObservedRunningTime="2025-11-24 17:14:43.219114559 +0000 UTC m=+841.378309618" Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.219931 4777 scope.go:117] "RemoveContainer" containerID="cbc3d8550cf41d4d9b2ce2194da843432ab6750e806ae46bf889432b690e0d57" Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.249405 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" podStartSLOduration=1.5050177 podStartE2EDuration="7.249378674s" podCreationTimestamp="2025-11-24 17:14:36 +0000 UTC" firstStartedPulling="2025-11-24 17:14:36.909543643 +0000 UTC m=+835.068738692" lastFinishedPulling="2025-11-24 17:14:42.653904617 +0000 UTC m=+840.813099666" observedRunningTime="2025-11-24 17:14:43.241412564 +0000 UTC m=+841.400607653" watchObservedRunningTime="2025-11-24 17:14:43.249378674 +0000 UTC m=+841.408573773" Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.268663 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4brxh"] Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.272473 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4brxh"] Nov 24 17:14:43 crc kubenswrapper[4777]: I1124 17:14:43.647992 4777 scope.go:117] "RemoveContainer" containerID="102ae36ed9ce8ddf96acc06d1168c28ad8ba207eb6d34f0b39e8483ebc33be32" Nov 24 17:14:45 crc kubenswrapper[4777]: I1124 17:14:45.256061 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0952d762-98cc-4c67-959a-772d127f6eb2" path="/var/lib/kubelet/pods/0952d762-98cc-4c67-959a-772d127f6eb2/volumes" Nov 24 17:14:47 crc kubenswrapper[4777]: I1124 17:14:47.131555 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:47 crc kubenswrapper[4777]: I1124 17:14:47.131623 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:47 crc kubenswrapper[4777]: I1124 17:14:47.199166 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:47 crc kubenswrapper[4777]: I1124 17:14:47.260167 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.205140 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xjmg9"] Nov 24 17:14:48 crc kubenswrapper[4777]: E1124 17:14:48.205651 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0952d762-98cc-4c67-959a-772d127f6eb2" containerName="extract-content" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.205665 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="0952d762-98cc-4c67-959a-772d127f6eb2" containerName="extract-content" Nov 24 17:14:48 crc kubenswrapper[4777]: E1124 17:14:48.205689 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0952d762-98cc-4c67-959a-772d127f6eb2" containerName="registry-server" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.205697 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="0952d762-98cc-4c67-959a-772d127f6eb2" containerName="registry-server" Nov 24 17:14:48 crc kubenswrapper[4777]: E1124 17:14:48.205710 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0952d762-98cc-4c67-959a-772d127f6eb2" containerName="extract-utilities" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.205718 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="0952d762-98cc-4c67-959a-772d127f6eb2" containerName="extract-utilities" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.205851 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="0952d762-98cc-4c67-959a-772d127f6eb2" containerName="registry-server" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.206783 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.229563 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xjmg9"] Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.262521 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25af89af-18fb-46a9-8190-8c515c163b54-catalog-content\") pod \"community-operators-xjmg9\" (UID: \"25af89af-18fb-46a9-8190-8c515c163b54\") " pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.262573 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25af89af-18fb-46a9-8190-8c515c163b54-utilities\") pod \"community-operators-xjmg9\" (UID: \"25af89af-18fb-46a9-8190-8c515c163b54\") " pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.262597 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xszr\" (UniqueName: \"kubernetes.io/projected/25af89af-18fb-46a9-8190-8c515c163b54-kube-api-access-4xszr\") pod \"community-operators-xjmg9\" (UID: \"25af89af-18fb-46a9-8190-8c515c163b54\") " pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.363286 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25af89af-18fb-46a9-8190-8c515c163b54-utilities\") pod \"community-operators-xjmg9\" (UID: \"25af89af-18fb-46a9-8190-8c515c163b54\") " pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.363346 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xszr\" (UniqueName: \"kubernetes.io/projected/25af89af-18fb-46a9-8190-8c515c163b54-kube-api-access-4xszr\") pod \"community-operators-xjmg9\" (UID: \"25af89af-18fb-46a9-8190-8c515c163b54\") " pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.363429 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25af89af-18fb-46a9-8190-8c515c163b54-catalog-content\") pod \"community-operators-xjmg9\" (UID: \"25af89af-18fb-46a9-8190-8c515c163b54\") " pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.363825 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25af89af-18fb-46a9-8190-8c515c163b54-utilities\") pod \"community-operators-xjmg9\" (UID: \"25af89af-18fb-46a9-8190-8c515c163b54\") " pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.363867 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25af89af-18fb-46a9-8190-8c515c163b54-catalog-content\") pod \"community-operators-xjmg9\" (UID: \"25af89af-18fb-46a9-8190-8c515c163b54\") " pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.387843 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xszr\" (UniqueName: \"kubernetes.io/projected/25af89af-18fb-46a9-8190-8c515c163b54-kube-api-access-4xszr\") pod \"community-operators-xjmg9\" (UID: \"25af89af-18fb-46a9-8190-8c515c163b54\") " pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:14:48 crc kubenswrapper[4777]: I1124 17:14:48.530626 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:14:49 crc kubenswrapper[4777]: I1124 17:14:49.059111 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xjmg9"] Nov 24 17:14:49 crc kubenswrapper[4777]: I1124 17:14:49.228415 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xjmg9" event={"ID":"25af89af-18fb-46a9-8190-8c515c163b54","Type":"ContainerStarted","Data":"527bfc3db16a89031162b8f855001d0033c176d16271c32ec1d986bc810f5e2d"} Nov 24 17:14:49 crc kubenswrapper[4777]: I1124 17:14:49.228478 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xjmg9" event={"ID":"25af89af-18fb-46a9-8190-8c515c163b54","Type":"ContainerStarted","Data":"d94d2e86b603adae5717bd50acf681230c013f91547e81ef4dde383be62dba25"} Nov 24 17:14:50 crc kubenswrapper[4777]: I1124 17:14:50.238108 4777 generic.go:334] "Generic (PLEG): container finished" podID="25af89af-18fb-46a9-8190-8c515c163b54" containerID="527bfc3db16a89031162b8f855001d0033c176d16271c32ec1d986bc810f5e2d" exitCode=0 Nov 24 17:14:50 crc kubenswrapper[4777]: I1124 17:14:50.238212 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xjmg9" event={"ID":"25af89af-18fb-46a9-8190-8c515c163b54","Type":"ContainerDied","Data":"527bfc3db16a89031162b8f855001d0033c176d16271c32ec1d986bc810f5e2d"} Nov 24 17:14:51 crc kubenswrapper[4777]: I1124 17:14:51.183033 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lt7fn"] Nov 24 17:14:51 crc kubenswrapper[4777]: I1124 17:14:51.183554 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lt7fn" podUID="75460ef2-43a2-4488-85eb-6f58ab746e0a" containerName="registry-server" containerID="cri-o://b440ba388414776d5107678ab7622d7f625ead43cd6c9868ee0d7624861d3caa" gracePeriod=2 Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.146735 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.251718 4777 generic.go:334] "Generic (PLEG): container finished" podID="75460ef2-43a2-4488-85eb-6f58ab746e0a" containerID="b440ba388414776d5107678ab7622d7f625ead43cd6c9868ee0d7624861d3caa" exitCode=0 Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.251765 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lt7fn" event={"ID":"75460ef2-43a2-4488-85eb-6f58ab746e0a","Type":"ContainerDied","Data":"b440ba388414776d5107678ab7622d7f625ead43cd6c9868ee0d7624861d3caa"} Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.251784 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lt7fn" event={"ID":"75460ef2-43a2-4488-85eb-6f58ab746e0a","Type":"ContainerDied","Data":"468498c1196b7eb45dbaf19b39e80d6f23bae6a9d48665d60f5325fa985b49fe"} Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.251800 4777 scope.go:117] "RemoveContainer" containerID="b440ba388414776d5107678ab7622d7f625ead43cd6c9868ee0d7624861d3caa" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.251878 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lt7fn" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.254453 4777 generic.go:334] "Generic (PLEG): container finished" podID="25af89af-18fb-46a9-8190-8c515c163b54" containerID="1b04ab343e74dcbe4c8897d70c8ef77f5eb276a01005d9ed7a01486e20b63c75" exitCode=0 Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.254475 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xjmg9" event={"ID":"25af89af-18fb-46a9-8190-8c515c163b54","Type":"ContainerDied","Data":"1b04ab343e74dcbe4c8897d70c8ef77f5eb276a01005d9ed7a01486e20b63c75"} Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.270753 4777 scope.go:117] "RemoveContainer" containerID="42cce577b6d8e8a3cf374c539beb4638e468583140d5974fa769f7e371c440f1" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.290151 4777 scope.go:117] "RemoveContainer" containerID="d713eda52aeff3d82ca51ffd43cdd191f319ab56306bf175018a1617712627fb" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.306935 4777 scope.go:117] "RemoveContainer" containerID="b440ba388414776d5107678ab7622d7f625ead43cd6c9868ee0d7624861d3caa" Nov 24 17:14:52 crc kubenswrapper[4777]: E1124 17:14:52.307272 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b440ba388414776d5107678ab7622d7f625ead43cd6c9868ee0d7624861d3caa\": container with ID starting with b440ba388414776d5107678ab7622d7f625ead43cd6c9868ee0d7624861d3caa not found: ID does not exist" containerID="b440ba388414776d5107678ab7622d7f625ead43cd6c9868ee0d7624861d3caa" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.307316 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b440ba388414776d5107678ab7622d7f625ead43cd6c9868ee0d7624861d3caa"} err="failed to get container status \"b440ba388414776d5107678ab7622d7f625ead43cd6c9868ee0d7624861d3caa\": rpc error: code = NotFound desc = could not find container \"b440ba388414776d5107678ab7622d7f625ead43cd6c9868ee0d7624861d3caa\": container with ID starting with b440ba388414776d5107678ab7622d7f625ead43cd6c9868ee0d7624861d3caa not found: ID does not exist" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.307335 4777 scope.go:117] "RemoveContainer" containerID="42cce577b6d8e8a3cf374c539beb4638e468583140d5974fa769f7e371c440f1" Nov 24 17:14:52 crc kubenswrapper[4777]: E1124 17:14:52.307650 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42cce577b6d8e8a3cf374c539beb4638e468583140d5974fa769f7e371c440f1\": container with ID starting with 42cce577b6d8e8a3cf374c539beb4638e468583140d5974fa769f7e371c440f1 not found: ID does not exist" containerID="42cce577b6d8e8a3cf374c539beb4638e468583140d5974fa769f7e371c440f1" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.307686 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42cce577b6d8e8a3cf374c539beb4638e468583140d5974fa769f7e371c440f1"} err="failed to get container status \"42cce577b6d8e8a3cf374c539beb4638e468583140d5974fa769f7e371c440f1\": rpc error: code = NotFound desc = could not find container \"42cce577b6d8e8a3cf374c539beb4638e468583140d5974fa769f7e371c440f1\": container with ID starting with 42cce577b6d8e8a3cf374c539beb4638e468583140d5974fa769f7e371c440f1 not found: ID does not exist" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.307716 4777 scope.go:117] "RemoveContainer" containerID="d713eda52aeff3d82ca51ffd43cdd191f319ab56306bf175018a1617712627fb" Nov 24 17:14:52 crc kubenswrapper[4777]: E1124 17:14:52.308038 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d713eda52aeff3d82ca51ffd43cdd191f319ab56306bf175018a1617712627fb\": container with ID starting with d713eda52aeff3d82ca51ffd43cdd191f319ab56306bf175018a1617712627fb not found: ID does not exist" containerID="d713eda52aeff3d82ca51ffd43cdd191f319ab56306bf175018a1617712627fb" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.308060 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d713eda52aeff3d82ca51ffd43cdd191f319ab56306bf175018a1617712627fb"} err="failed to get container status \"d713eda52aeff3d82ca51ffd43cdd191f319ab56306bf175018a1617712627fb\": rpc error: code = NotFound desc = could not find container \"d713eda52aeff3d82ca51ffd43cdd191f319ab56306bf175018a1617712627fb\": container with ID starting with d713eda52aeff3d82ca51ffd43cdd191f319ab56306bf175018a1617712627fb not found: ID does not exist" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.319667 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75460ef2-43a2-4488-85eb-6f58ab746e0a-utilities\") pod \"75460ef2-43a2-4488-85eb-6f58ab746e0a\" (UID: \"75460ef2-43a2-4488-85eb-6f58ab746e0a\") " Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.319717 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7mwd\" (UniqueName: \"kubernetes.io/projected/75460ef2-43a2-4488-85eb-6f58ab746e0a-kube-api-access-m7mwd\") pod \"75460ef2-43a2-4488-85eb-6f58ab746e0a\" (UID: \"75460ef2-43a2-4488-85eb-6f58ab746e0a\") " Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.319987 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75460ef2-43a2-4488-85eb-6f58ab746e0a-catalog-content\") pod \"75460ef2-43a2-4488-85eb-6f58ab746e0a\" (UID: \"75460ef2-43a2-4488-85eb-6f58ab746e0a\") " Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.321034 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75460ef2-43a2-4488-85eb-6f58ab746e0a-utilities" (OuterVolumeSpecName: "utilities") pod "75460ef2-43a2-4488-85eb-6f58ab746e0a" (UID: "75460ef2-43a2-4488-85eb-6f58ab746e0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.321178 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75460ef2-43a2-4488-85eb-6f58ab746e0a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.325360 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75460ef2-43a2-4488-85eb-6f58ab746e0a-kube-api-access-m7mwd" (OuterVolumeSpecName: "kube-api-access-m7mwd") pod "75460ef2-43a2-4488-85eb-6f58ab746e0a" (UID: "75460ef2-43a2-4488-85eb-6f58ab746e0a"). InnerVolumeSpecName "kube-api-access-m7mwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.341050 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75460ef2-43a2-4488-85eb-6f58ab746e0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75460ef2-43a2-4488-85eb-6f58ab746e0a" (UID: "75460ef2-43a2-4488-85eb-6f58ab746e0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.422899 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7mwd\" (UniqueName: \"kubernetes.io/projected/75460ef2-43a2-4488-85eb-6f58ab746e0a-kube-api-access-m7mwd\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.422937 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75460ef2-43a2-4488-85eb-6f58ab746e0a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.576717 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lt7fn"] Nov 24 17:14:52 crc kubenswrapper[4777]: I1124 17:14:52.580310 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lt7fn"] Nov 24 17:14:53 crc kubenswrapper[4777]: I1124 17:14:53.256185 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75460ef2-43a2-4488-85eb-6f58ab746e0a" path="/var/lib/kubelet/pods/75460ef2-43a2-4488-85eb-6f58ab746e0a/volumes" Nov 24 17:14:53 crc kubenswrapper[4777]: I1124 17:14:53.263558 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xjmg9" event={"ID":"25af89af-18fb-46a9-8190-8c515c163b54","Type":"ContainerStarted","Data":"4565b2bb4d14eb8fbd1774e0b3e4fe8f73d94b1ab2473d3dd98ae3fa68afac2d"} Nov 24 17:14:53 crc kubenswrapper[4777]: I1124 17:14:53.289151 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xjmg9" podStartSLOduration=2.877620292 podStartE2EDuration="5.289123861s" podCreationTimestamp="2025-11-24 17:14:48 +0000 UTC" firstStartedPulling="2025-11-24 17:14:50.24123823 +0000 UTC m=+848.400433319" lastFinishedPulling="2025-11-24 17:14:52.652741829 +0000 UTC m=+850.811936888" observedRunningTime="2025-11-24 17:14:53.284893479 +0000 UTC m=+851.444088528" watchObservedRunningTime="2025-11-24 17:14:53.289123861 +0000 UTC m=+851.448318930" Nov 24 17:14:56 crc kubenswrapper[4777]: I1124 17:14:56.660855 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7578594d7b-kn9dd" Nov 24 17:14:58 crc kubenswrapper[4777]: I1124 17:14:58.531473 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:14:58 crc kubenswrapper[4777]: I1124 17:14:58.531539 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:14:58 crc kubenswrapper[4777]: I1124 17:14:58.613209 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:14:59 crc kubenswrapper[4777]: I1124 17:14:59.357655 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.129208 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb"] Nov 24 17:15:00 crc kubenswrapper[4777]: E1124 17:15:00.129450 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75460ef2-43a2-4488-85eb-6f58ab746e0a" containerName="registry-server" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.129463 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="75460ef2-43a2-4488-85eb-6f58ab746e0a" containerName="registry-server" Nov 24 17:15:00 crc kubenswrapper[4777]: E1124 17:15:00.129479 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75460ef2-43a2-4488-85eb-6f58ab746e0a" containerName="extract-utilities" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.129486 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="75460ef2-43a2-4488-85eb-6f58ab746e0a" containerName="extract-utilities" Nov 24 17:15:00 crc kubenswrapper[4777]: E1124 17:15:00.129497 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75460ef2-43a2-4488-85eb-6f58ab746e0a" containerName="extract-content" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.129503 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="75460ef2-43a2-4488-85eb-6f58ab746e0a" containerName="extract-content" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.129610 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="75460ef2-43a2-4488-85eb-6f58ab746e0a" containerName="registry-server" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.129994 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.132127 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.132476 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.173632 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb"] Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.185870 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xjmg9"] Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.235798 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/631cfc31-4721-408b-860c-e7675ed55892-config-volume\") pod \"collect-profiles-29400075-rppzb\" (UID: \"631cfc31-4721-408b-860c-e7675ed55892\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.235857 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knw9n\" (UniqueName: \"kubernetes.io/projected/631cfc31-4721-408b-860c-e7675ed55892-kube-api-access-knw9n\") pod \"collect-profiles-29400075-rppzb\" (UID: \"631cfc31-4721-408b-860c-e7675ed55892\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.235936 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/631cfc31-4721-408b-860c-e7675ed55892-secret-volume\") pod \"collect-profiles-29400075-rppzb\" (UID: \"631cfc31-4721-408b-860c-e7675ed55892\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.337455 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/631cfc31-4721-408b-860c-e7675ed55892-secret-volume\") pod \"collect-profiles-29400075-rppzb\" (UID: \"631cfc31-4721-408b-860c-e7675ed55892\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.337603 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/631cfc31-4721-408b-860c-e7675ed55892-config-volume\") pod \"collect-profiles-29400075-rppzb\" (UID: \"631cfc31-4721-408b-860c-e7675ed55892\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.337662 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knw9n\" (UniqueName: \"kubernetes.io/projected/631cfc31-4721-408b-860c-e7675ed55892-kube-api-access-knw9n\") pod \"collect-profiles-29400075-rppzb\" (UID: \"631cfc31-4721-408b-860c-e7675ed55892\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.339201 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/631cfc31-4721-408b-860c-e7675ed55892-config-volume\") pod \"collect-profiles-29400075-rppzb\" (UID: \"631cfc31-4721-408b-860c-e7675ed55892\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.346802 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/631cfc31-4721-408b-860c-e7675ed55892-secret-volume\") pod \"collect-profiles-29400075-rppzb\" (UID: \"631cfc31-4721-408b-860c-e7675ed55892\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.357196 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knw9n\" (UniqueName: \"kubernetes.io/projected/631cfc31-4721-408b-860c-e7675ed55892-kube-api-access-knw9n\") pod \"collect-profiles-29400075-rppzb\" (UID: \"631cfc31-4721-408b-860c-e7675ed55892\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.448334 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" Nov 24 17:15:00 crc kubenswrapper[4777]: I1124 17:15:00.964884 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb"] Nov 24 17:15:01 crc kubenswrapper[4777]: I1124 17:15:01.319049 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" event={"ID":"631cfc31-4721-408b-860c-e7675ed55892","Type":"ContainerStarted","Data":"2b2cef017dfeb04eb5a9c9c3dcd96dca5d5ba348ce770e2822ac356f37ca87b2"} Nov 24 17:15:01 crc kubenswrapper[4777]: I1124 17:15:01.319208 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xjmg9" podUID="25af89af-18fb-46a9-8190-8c515c163b54" containerName="registry-server" containerID="cri-o://4565b2bb4d14eb8fbd1774e0b3e4fe8f73d94b1ab2473d3dd98ae3fa68afac2d" gracePeriod=2 Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.305037 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.329505 4777 generic.go:334] "Generic (PLEG): container finished" podID="631cfc31-4721-408b-860c-e7675ed55892" containerID="ec4843ab252ebb79bcc5b433b19d8e6a94adb0eebb269d8a732c4edf6fbf5120" exitCode=0 Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.329567 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" event={"ID":"631cfc31-4721-408b-860c-e7675ed55892","Type":"ContainerDied","Data":"ec4843ab252ebb79bcc5b433b19d8e6a94adb0eebb269d8a732c4edf6fbf5120"} Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.331954 4777 generic.go:334] "Generic (PLEG): container finished" podID="25af89af-18fb-46a9-8190-8c515c163b54" containerID="4565b2bb4d14eb8fbd1774e0b3e4fe8f73d94b1ab2473d3dd98ae3fa68afac2d" exitCode=0 Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.332009 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xjmg9" event={"ID":"25af89af-18fb-46a9-8190-8c515c163b54","Type":"ContainerDied","Data":"4565b2bb4d14eb8fbd1774e0b3e4fe8f73d94b1ab2473d3dd98ae3fa68afac2d"} Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.332036 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xjmg9" event={"ID":"25af89af-18fb-46a9-8190-8c515c163b54","Type":"ContainerDied","Data":"d94d2e86b603adae5717bd50acf681230c013f91547e81ef4dde383be62dba25"} Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.332056 4777 scope.go:117] "RemoveContainer" containerID="4565b2bb4d14eb8fbd1774e0b3e4fe8f73d94b1ab2473d3dd98ae3fa68afac2d" Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.332499 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xjmg9" Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.349457 4777 scope.go:117] "RemoveContainer" containerID="1b04ab343e74dcbe4c8897d70c8ef77f5eb276a01005d9ed7a01486e20b63c75" Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.365692 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25af89af-18fb-46a9-8190-8c515c163b54-utilities\") pod \"25af89af-18fb-46a9-8190-8c515c163b54\" (UID: \"25af89af-18fb-46a9-8190-8c515c163b54\") " Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.365755 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25af89af-18fb-46a9-8190-8c515c163b54-catalog-content\") pod \"25af89af-18fb-46a9-8190-8c515c163b54\" (UID: \"25af89af-18fb-46a9-8190-8c515c163b54\") " Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.365849 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xszr\" (UniqueName: \"kubernetes.io/projected/25af89af-18fb-46a9-8190-8c515c163b54-kube-api-access-4xszr\") pod \"25af89af-18fb-46a9-8190-8c515c163b54\" (UID: \"25af89af-18fb-46a9-8190-8c515c163b54\") " Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.368232 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25af89af-18fb-46a9-8190-8c515c163b54-utilities" (OuterVolumeSpecName: "utilities") pod "25af89af-18fb-46a9-8190-8c515c163b54" (UID: "25af89af-18fb-46a9-8190-8c515c163b54"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.375015 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25af89af-18fb-46a9-8190-8c515c163b54-kube-api-access-4xszr" (OuterVolumeSpecName: "kube-api-access-4xszr") pod "25af89af-18fb-46a9-8190-8c515c163b54" (UID: "25af89af-18fb-46a9-8190-8c515c163b54"). InnerVolumeSpecName "kube-api-access-4xszr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.395568 4777 scope.go:117] "RemoveContainer" containerID="527bfc3db16a89031162b8f855001d0033c176d16271c32ec1d986bc810f5e2d" Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.448906 4777 scope.go:117] "RemoveContainer" containerID="4565b2bb4d14eb8fbd1774e0b3e4fe8f73d94b1ab2473d3dd98ae3fa68afac2d" Nov 24 17:15:02 crc kubenswrapper[4777]: E1124 17:15:02.451375 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4565b2bb4d14eb8fbd1774e0b3e4fe8f73d94b1ab2473d3dd98ae3fa68afac2d\": container with ID starting with 4565b2bb4d14eb8fbd1774e0b3e4fe8f73d94b1ab2473d3dd98ae3fa68afac2d not found: ID does not exist" containerID="4565b2bb4d14eb8fbd1774e0b3e4fe8f73d94b1ab2473d3dd98ae3fa68afac2d" Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.451412 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4565b2bb4d14eb8fbd1774e0b3e4fe8f73d94b1ab2473d3dd98ae3fa68afac2d"} err="failed to get container status \"4565b2bb4d14eb8fbd1774e0b3e4fe8f73d94b1ab2473d3dd98ae3fa68afac2d\": rpc error: code = NotFound desc = could not find container \"4565b2bb4d14eb8fbd1774e0b3e4fe8f73d94b1ab2473d3dd98ae3fa68afac2d\": container with ID starting with 4565b2bb4d14eb8fbd1774e0b3e4fe8f73d94b1ab2473d3dd98ae3fa68afac2d not found: ID does not exist" Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.451438 4777 scope.go:117] "RemoveContainer" containerID="1b04ab343e74dcbe4c8897d70c8ef77f5eb276a01005d9ed7a01486e20b63c75" Nov 24 17:15:02 crc kubenswrapper[4777]: E1124 17:15:02.451884 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b04ab343e74dcbe4c8897d70c8ef77f5eb276a01005d9ed7a01486e20b63c75\": container with ID starting with 1b04ab343e74dcbe4c8897d70c8ef77f5eb276a01005d9ed7a01486e20b63c75 not found: ID does not exist" containerID="1b04ab343e74dcbe4c8897d70c8ef77f5eb276a01005d9ed7a01486e20b63c75" Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.451945 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b04ab343e74dcbe4c8897d70c8ef77f5eb276a01005d9ed7a01486e20b63c75"} err="failed to get container status \"1b04ab343e74dcbe4c8897d70c8ef77f5eb276a01005d9ed7a01486e20b63c75\": rpc error: code = NotFound desc = could not find container \"1b04ab343e74dcbe4c8897d70c8ef77f5eb276a01005d9ed7a01486e20b63c75\": container with ID starting with 1b04ab343e74dcbe4c8897d70c8ef77f5eb276a01005d9ed7a01486e20b63c75 not found: ID does not exist" Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.451994 4777 scope.go:117] "RemoveContainer" containerID="527bfc3db16a89031162b8f855001d0033c176d16271c32ec1d986bc810f5e2d" Nov 24 17:15:02 crc kubenswrapper[4777]: E1124 17:15:02.452640 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"527bfc3db16a89031162b8f855001d0033c176d16271c32ec1d986bc810f5e2d\": container with ID starting with 527bfc3db16a89031162b8f855001d0033c176d16271c32ec1d986bc810f5e2d not found: ID does not exist" containerID="527bfc3db16a89031162b8f855001d0033c176d16271c32ec1d986bc810f5e2d" Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.452670 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"527bfc3db16a89031162b8f855001d0033c176d16271c32ec1d986bc810f5e2d"} err="failed to get container status \"527bfc3db16a89031162b8f855001d0033c176d16271c32ec1d986bc810f5e2d\": rpc error: code = NotFound desc = could not find container \"527bfc3db16a89031162b8f855001d0033c176d16271c32ec1d986bc810f5e2d\": container with ID starting with 527bfc3db16a89031162b8f855001d0033c176d16271c32ec1d986bc810f5e2d not found: ID does not exist" Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.467745 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xszr\" (UniqueName: \"kubernetes.io/projected/25af89af-18fb-46a9-8190-8c515c163b54-kube-api-access-4xszr\") on node \"crc\" DevicePath \"\"" Nov 24 17:15:02 crc kubenswrapper[4777]: I1124 17:15:02.467769 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25af89af-18fb-46a9-8190-8c515c163b54-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:15:03 crc kubenswrapper[4777]: I1124 17:15:03.189525 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25af89af-18fb-46a9-8190-8c515c163b54-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25af89af-18fb-46a9-8190-8c515c163b54" (UID: "25af89af-18fb-46a9-8190-8c515c163b54"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:15:03 crc kubenswrapper[4777]: I1124 17:15:03.272641 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xjmg9"] Nov 24 17:15:03 crc kubenswrapper[4777]: I1124 17:15:03.276095 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xjmg9"] Nov 24 17:15:03 crc kubenswrapper[4777]: I1124 17:15:03.280558 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25af89af-18fb-46a9-8190-8c515c163b54-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:15:03 crc kubenswrapper[4777]: I1124 17:15:03.670724 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" Nov 24 17:15:03 crc kubenswrapper[4777]: I1124 17:15:03.685288 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knw9n\" (UniqueName: \"kubernetes.io/projected/631cfc31-4721-408b-860c-e7675ed55892-kube-api-access-knw9n\") pod \"631cfc31-4721-408b-860c-e7675ed55892\" (UID: \"631cfc31-4721-408b-860c-e7675ed55892\") " Nov 24 17:15:03 crc kubenswrapper[4777]: I1124 17:15:03.685373 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/631cfc31-4721-408b-860c-e7675ed55892-secret-volume\") pod \"631cfc31-4721-408b-860c-e7675ed55892\" (UID: \"631cfc31-4721-408b-860c-e7675ed55892\") " Nov 24 17:15:03 crc kubenswrapper[4777]: I1124 17:15:03.685414 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/631cfc31-4721-408b-860c-e7675ed55892-config-volume\") pod \"631cfc31-4721-408b-860c-e7675ed55892\" (UID: \"631cfc31-4721-408b-860c-e7675ed55892\") " Nov 24 17:15:03 crc kubenswrapper[4777]: I1124 17:15:03.686302 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/631cfc31-4721-408b-860c-e7675ed55892-config-volume" (OuterVolumeSpecName: "config-volume") pod "631cfc31-4721-408b-860c-e7675ed55892" (UID: "631cfc31-4721-408b-860c-e7675ed55892"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:15:03 crc kubenswrapper[4777]: I1124 17:15:03.691196 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/631cfc31-4721-408b-860c-e7675ed55892-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "631cfc31-4721-408b-860c-e7675ed55892" (UID: "631cfc31-4721-408b-860c-e7675ed55892"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:15:03 crc kubenswrapper[4777]: I1124 17:15:03.693403 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/631cfc31-4721-408b-860c-e7675ed55892-kube-api-access-knw9n" (OuterVolumeSpecName: "kube-api-access-knw9n") pod "631cfc31-4721-408b-860c-e7675ed55892" (UID: "631cfc31-4721-408b-860c-e7675ed55892"). InnerVolumeSpecName "kube-api-access-knw9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:15:03 crc kubenswrapper[4777]: I1124 17:15:03.786491 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knw9n\" (UniqueName: \"kubernetes.io/projected/631cfc31-4721-408b-860c-e7675ed55892-kube-api-access-knw9n\") on node \"crc\" DevicePath \"\"" Nov 24 17:15:03 crc kubenswrapper[4777]: I1124 17:15:03.786520 4777 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/631cfc31-4721-408b-860c-e7675ed55892-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:15:03 crc kubenswrapper[4777]: I1124 17:15:03.786529 4777 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/631cfc31-4721-408b-860c-e7675ed55892-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:15:04 crc kubenswrapper[4777]: I1124 17:15:04.349631 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" event={"ID":"631cfc31-4721-408b-860c-e7675ed55892","Type":"ContainerDied","Data":"2b2cef017dfeb04eb5a9c9c3dcd96dca5d5ba348ce770e2822ac356f37ca87b2"} Nov 24 17:15:04 crc kubenswrapper[4777]: I1124 17:15:04.349683 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b2cef017dfeb04eb5a9c9c3dcd96dca5d5ba348ce770e2822ac356f37ca87b2" Nov 24 17:15:04 crc kubenswrapper[4777]: I1124 17:15:04.349696 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb" Nov 24 17:15:05 crc kubenswrapper[4777]: I1124 17:15:05.258551 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25af89af-18fb-46a9-8190-8c515c163b54" path="/var/lib/kubelet/pods/25af89af-18fb-46a9-8190-8c515c163b54/volumes" Nov 24 17:15:14 crc kubenswrapper[4777]: I1124 17:15:14.270808 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:15:14 crc kubenswrapper[4777]: I1124 17:15:14.271521 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:15:16 crc kubenswrapper[4777]: I1124 17:15:16.423822 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7444c76584-h5nzb" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.177202 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-c95bf"] Nov 24 17:15:17 crc kubenswrapper[4777]: E1124 17:15:17.177650 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631cfc31-4721-408b-860c-e7675ed55892" containerName="collect-profiles" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.177663 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="631cfc31-4721-408b-860c-e7675ed55892" containerName="collect-profiles" Nov 24 17:15:17 crc kubenswrapper[4777]: E1124 17:15:17.177679 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25af89af-18fb-46a9-8190-8c515c163b54" containerName="registry-server" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.177686 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="25af89af-18fb-46a9-8190-8c515c163b54" containerName="registry-server" Nov 24 17:15:17 crc kubenswrapper[4777]: E1124 17:15:17.177696 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25af89af-18fb-46a9-8190-8c515c163b54" containerName="extract-content" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.177702 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="25af89af-18fb-46a9-8190-8c515c163b54" containerName="extract-content" Nov 24 17:15:17 crc kubenswrapper[4777]: E1124 17:15:17.177717 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25af89af-18fb-46a9-8190-8c515c163b54" containerName="extract-utilities" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.177723 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="25af89af-18fb-46a9-8190-8c515c163b54" containerName="extract-utilities" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.177831 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="25af89af-18fb-46a9-8190-8c515c163b54" containerName="registry-server" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.177844 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="631cfc31-4721-408b-860c-e7675ed55892" containerName="collect-profiles" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.179719 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.182227 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.182229 4777 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-8twrv" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.182528 4777 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.184095 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-86vqd"] Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.185394 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-86vqd" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.187508 4777 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.202680 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-86vqd"] Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.230694 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-frr-sockets\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.230743 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-metrics-certs\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.230791 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a-cert\") pod \"frr-k8s-webhook-server-6998585d5-86vqd\" (UID: \"2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-86vqd" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.230848 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dskxt\" (UniqueName: \"kubernetes.io/projected/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-kube-api-access-dskxt\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.230872 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tpjt\" (UniqueName: \"kubernetes.io/projected/2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a-kube-api-access-2tpjt\") pod \"frr-k8s-webhook-server-6998585d5-86vqd\" (UID: \"2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-86vqd" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.230941 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-frr-startup\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.230985 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-frr-conf\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.231029 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-reloader\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.231085 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-metrics\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.263750 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-pz9ts"] Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.264904 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-pz9ts" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.267801 4777 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.268011 4777 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.268162 4777 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-qzbvc" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.269831 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.275576 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-z9z24"] Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.276413 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-z9z24" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.277569 4777 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.291601 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-z9z24"] Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332198 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a-cert\") pod \"frr-k8s-webhook-server-6998585d5-86vqd\" (UID: \"2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-86vqd" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332243 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5ae956fa-5a45-4248-b300-d802aa171d52-memberlist\") pod \"speaker-pz9ts\" (UID: \"5ae956fa-5a45-4248-b300-d802aa171d52\") " pod="metallb-system/speaker-pz9ts" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332271 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b4a78161-cf5f-4307-9adb-a05de51d620f-cert\") pod \"controller-6c7b4b5f48-z9z24\" (UID: \"b4a78161-cf5f-4307-9adb-a05de51d620f\") " pod="metallb-system/controller-6c7b4b5f48-z9z24" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332359 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dskxt\" (UniqueName: \"kubernetes.io/projected/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-kube-api-access-dskxt\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332379 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tpjt\" (UniqueName: \"kubernetes.io/projected/2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a-kube-api-access-2tpjt\") pod \"frr-k8s-webhook-server-6998585d5-86vqd\" (UID: \"2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-86vqd" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332416 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-frr-startup\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332436 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5ae956fa-5a45-4248-b300-d802aa171d52-metrics-certs\") pod \"speaker-pz9ts\" (UID: \"5ae956fa-5a45-4248-b300-d802aa171d52\") " pod="metallb-system/speaker-pz9ts" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332463 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-frr-conf\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332498 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbxbs\" (UniqueName: \"kubernetes.io/projected/5ae956fa-5a45-4248-b300-d802aa171d52-kube-api-access-tbxbs\") pod \"speaker-pz9ts\" (UID: \"5ae956fa-5a45-4248-b300-d802aa171d52\") " pod="metallb-system/speaker-pz9ts" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332528 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5ae956fa-5a45-4248-b300-d802aa171d52-metallb-excludel2\") pod \"speaker-pz9ts\" (UID: \"5ae956fa-5a45-4248-b300-d802aa171d52\") " pod="metallb-system/speaker-pz9ts" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332549 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-reloader\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332579 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-metrics\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332602 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97cgr\" (UniqueName: \"kubernetes.io/projected/b4a78161-cf5f-4307-9adb-a05de51d620f-kube-api-access-97cgr\") pod \"controller-6c7b4b5f48-z9z24\" (UID: \"b4a78161-cf5f-4307-9adb-a05de51d620f\") " pod="metallb-system/controller-6c7b4b5f48-z9z24" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332643 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-frr-sockets\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332667 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-metrics-certs\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.332700 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4a78161-cf5f-4307-9adb-a05de51d620f-metrics-certs\") pod \"controller-6c7b4b5f48-z9z24\" (UID: \"b4a78161-cf5f-4307-9adb-a05de51d620f\") " pod="metallb-system/controller-6c7b4b5f48-z9z24" Nov 24 17:15:17 crc kubenswrapper[4777]: E1124 17:15:17.332834 4777 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 24 17:15:17 crc kubenswrapper[4777]: E1124 17:15:17.332878 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a-cert podName:2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a nodeName:}" failed. No retries permitted until 2025-11-24 17:15:17.832861417 +0000 UTC m=+875.992056466 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a-cert") pod "frr-k8s-webhook-server-6998585d5-86vqd" (UID: "2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a") : secret "frr-k8s-webhook-server-cert" not found Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.333414 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-frr-conf\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.333478 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-metrics\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.333606 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-reloader\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.333786 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-frr-sockets\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.333921 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-frr-startup\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.342810 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-metrics-certs\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.348212 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dskxt\" (UniqueName: \"kubernetes.io/projected/0dd8e42b-e649-438d-bd92-e35a3ed06d4a-kube-api-access-dskxt\") pod \"frr-k8s-c95bf\" (UID: \"0dd8e42b-e649-438d-bd92-e35a3ed06d4a\") " pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.351305 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tpjt\" (UniqueName: \"kubernetes.io/projected/2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a-kube-api-access-2tpjt\") pod \"frr-k8s-webhook-server-6998585d5-86vqd\" (UID: \"2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-86vqd" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.433527 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbxbs\" (UniqueName: \"kubernetes.io/projected/5ae956fa-5a45-4248-b300-d802aa171d52-kube-api-access-tbxbs\") pod \"speaker-pz9ts\" (UID: \"5ae956fa-5a45-4248-b300-d802aa171d52\") " pod="metallb-system/speaker-pz9ts" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.433576 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5ae956fa-5a45-4248-b300-d802aa171d52-metallb-excludel2\") pod \"speaker-pz9ts\" (UID: \"5ae956fa-5a45-4248-b300-d802aa171d52\") " pod="metallb-system/speaker-pz9ts" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.433608 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97cgr\" (UniqueName: \"kubernetes.io/projected/b4a78161-cf5f-4307-9adb-a05de51d620f-kube-api-access-97cgr\") pod \"controller-6c7b4b5f48-z9z24\" (UID: \"b4a78161-cf5f-4307-9adb-a05de51d620f\") " pod="metallb-system/controller-6c7b4b5f48-z9z24" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.433638 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4a78161-cf5f-4307-9adb-a05de51d620f-metrics-certs\") pod \"controller-6c7b4b5f48-z9z24\" (UID: \"b4a78161-cf5f-4307-9adb-a05de51d620f\") " pod="metallb-system/controller-6c7b4b5f48-z9z24" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.433682 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5ae956fa-5a45-4248-b300-d802aa171d52-memberlist\") pod \"speaker-pz9ts\" (UID: \"5ae956fa-5a45-4248-b300-d802aa171d52\") " pod="metallb-system/speaker-pz9ts" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.433707 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b4a78161-cf5f-4307-9adb-a05de51d620f-cert\") pod \"controller-6c7b4b5f48-z9z24\" (UID: \"b4a78161-cf5f-4307-9adb-a05de51d620f\") " pod="metallb-system/controller-6c7b4b5f48-z9z24" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.433762 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5ae956fa-5a45-4248-b300-d802aa171d52-metrics-certs\") pod \"speaker-pz9ts\" (UID: \"5ae956fa-5a45-4248-b300-d802aa171d52\") " pod="metallb-system/speaker-pz9ts" Nov 24 17:15:17 crc kubenswrapper[4777]: E1124 17:15:17.433789 4777 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 17:15:17 crc kubenswrapper[4777]: E1124 17:15:17.433808 4777 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Nov 24 17:15:17 crc kubenswrapper[4777]: E1124 17:15:17.433848 4777 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Nov 24 17:15:17 crc kubenswrapper[4777]: E1124 17:15:17.433850 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5ae956fa-5a45-4248-b300-d802aa171d52-memberlist podName:5ae956fa-5a45-4248-b300-d802aa171d52 nodeName:}" failed. No retries permitted until 2025-11-24 17:15:17.933834578 +0000 UTC m=+876.093029627 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/5ae956fa-5a45-4248-b300-d802aa171d52-memberlist") pod "speaker-pz9ts" (UID: "5ae956fa-5a45-4248-b300-d802aa171d52") : secret "metallb-memberlist" not found Nov 24 17:15:17 crc kubenswrapper[4777]: E1124 17:15:17.433901 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b4a78161-cf5f-4307-9adb-a05de51d620f-metrics-certs podName:b4a78161-cf5f-4307-9adb-a05de51d620f nodeName:}" failed. No retries permitted until 2025-11-24 17:15:17.93388616 +0000 UTC m=+876.093081319 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b4a78161-cf5f-4307-9adb-a05de51d620f-metrics-certs") pod "controller-6c7b4b5f48-z9z24" (UID: "b4a78161-cf5f-4307-9adb-a05de51d620f") : secret "controller-certs-secret" not found Nov 24 17:15:17 crc kubenswrapper[4777]: E1124 17:15:17.433914 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5ae956fa-5a45-4248-b300-d802aa171d52-metrics-certs podName:5ae956fa-5a45-4248-b300-d802aa171d52 nodeName:}" failed. No retries permitted until 2025-11-24 17:15:17.933908451 +0000 UTC m=+876.093103500 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5ae956fa-5a45-4248-b300-d802aa171d52-metrics-certs") pod "speaker-pz9ts" (UID: "5ae956fa-5a45-4248-b300-d802aa171d52") : secret "speaker-certs-secret" not found Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.434329 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5ae956fa-5a45-4248-b300-d802aa171d52-metallb-excludel2\") pod \"speaker-pz9ts\" (UID: \"5ae956fa-5a45-4248-b300-d802aa171d52\") " pod="metallb-system/speaker-pz9ts" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.436048 4777 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.448696 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b4a78161-cf5f-4307-9adb-a05de51d620f-cert\") pod \"controller-6c7b4b5f48-z9z24\" (UID: \"b4a78161-cf5f-4307-9adb-a05de51d620f\") " pod="metallb-system/controller-6c7b4b5f48-z9z24" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.451837 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbxbs\" (UniqueName: \"kubernetes.io/projected/5ae956fa-5a45-4248-b300-d802aa171d52-kube-api-access-tbxbs\") pod \"speaker-pz9ts\" (UID: \"5ae956fa-5a45-4248-b300-d802aa171d52\") " pod="metallb-system/speaker-pz9ts" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.452511 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97cgr\" (UniqueName: \"kubernetes.io/projected/b4a78161-cf5f-4307-9adb-a05de51d620f-kube-api-access-97cgr\") pod \"controller-6c7b4b5f48-z9z24\" (UID: \"b4a78161-cf5f-4307-9adb-a05de51d620f\") " pod="metallb-system/controller-6c7b4b5f48-z9z24" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.504034 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.838996 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a-cert\") pod \"frr-k8s-webhook-server-6998585d5-86vqd\" (UID: \"2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-86vqd" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.845332 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a-cert\") pod \"frr-k8s-webhook-server-6998585d5-86vqd\" (UID: \"2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-86vqd" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.940941 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5ae956fa-5a45-4248-b300-d802aa171d52-memberlist\") pod \"speaker-pz9ts\" (UID: \"5ae956fa-5a45-4248-b300-d802aa171d52\") " pod="metallb-system/speaker-pz9ts" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.941175 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5ae956fa-5a45-4248-b300-d802aa171d52-metrics-certs\") pod \"speaker-pz9ts\" (UID: \"5ae956fa-5a45-4248-b300-d802aa171d52\") " pod="metallb-system/speaker-pz9ts" Nov 24 17:15:17 crc kubenswrapper[4777]: E1124 17:15:17.941201 4777 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 17:15:17 crc kubenswrapper[4777]: E1124 17:15:17.941295 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5ae956fa-5a45-4248-b300-d802aa171d52-memberlist podName:5ae956fa-5a45-4248-b300-d802aa171d52 nodeName:}" failed. No retries permitted until 2025-11-24 17:15:18.94126786 +0000 UTC m=+877.100462949 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/5ae956fa-5a45-4248-b300-d802aa171d52-memberlist") pod "speaker-pz9ts" (UID: "5ae956fa-5a45-4248-b300-d802aa171d52") : secret "metallb-memberlist" not found Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.941336 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4a78161-cf5f-4307-9adb-a05de51d620f-metrics-certs\") pod \"controller-6c7b4b5f48-z9z24\" (UID: \"b4a78161-cf5f-4307-9adb-a05de51d620f\") " pod="metallb-system/controller-6c7b4b5f48-z9z24" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.946426 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5ae956fa-5a45-4248-b300-d802aa171d52-metrics-certs\") pod \"speaker-pz9ts\" (UID: \"5ae956fa-5a45-4248-b300-d802aa171d52\") " pod="metallb-system/speaker-pz9ts" Nov 24 17:15:17 crc kubenswrapper[4777]: I1124 17:15:17.946928 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4a78161-cf5f-4307-9adb-a05de51d620f-metrics-certs\") pod \"controller-6c7b4b5f48-z9z24\" (UID: \"b4a78161-cf5f-4307-9adb-a05de51d620f\") " pod="metallb-system/controller-6c7b4b5f48-z9z24" Nov 24 17:15:18 crc kubenswrapper[4777]: I1124 17:15:18.115948 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-86vqd" Nov 24 17:15:18 crc kubenswrapper[4777]: I1124 17:15:18.200207 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-z9z24" Nov 24 17:15:18 crc kubenswrapper[4777]: I1124 17:15:18.418344 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-86vqd"] Nov 24 17:15:18 crc kubenswrapper[4777]: W1124 17:15:18.423557 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e13d83f_3c71_4981_9cbb_9d0ef94c1a5a.slice/crio-3ec969222cb78d65f59c7d9d4101a0d9e1c7be9a00473515194c2becf23cf414 WatchSource:0}: Error finding container 3ec969222cb78d65f59c7d9d4101a0d9e1c7be9a00473515194c2becf23cf414: Status 404 returned error can't find the container with id 3ec969222cb78d65f59c7d9d4101a0d9e1c7be9a00473515194c2becf23cf414 Nov 24 17:15:18 crc kubenswrapper[4777]: I1124 17:15:18.451660 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95bf" event={"ID":"0dd8e42b-e649-438d-bd92-e35a3ed06d4a","Type":"ContainerStarted","Data":"68f421bba0e4b76d63f2f38e510e2faa127c111694160699dd997d1e5c7bfa9e"} Nov 24 17:15:18 crc kubenswrapper[4777]: I1124 17:15:18.453039 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-86vqd" event={"ID":"2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a","Type":"ContainerStarted","Data":"3ec969222cb78d65f59c7d9d4101a0d9e1c7be9a00473515194c2becf23cf414"} Nov 24 17:15:18 crc kubenswrapper[4777]: I1124 17:15:18.675043 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-z9z24"] Nov 24 17:15:18 crc kubenswrapper[4777]: I1124 17:15:18.953716 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5ae956fa-5a45-4248-b300-d802aa171d52-memberlist\") pod \"speaker-pz9ts\" (UID: \"5ae956fa-5a45-4248-b300-d802aa171d52\") " pod="metallb-system/speaker-pz9ts" Nov 24 17:15:18 crc kubenswrapper[4777]: I1124 17:15:18.959630 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5ae956fa-5a45-4248-b300-d802aa171d52-memberlist\") pod \"speaker-pz9ts\" (UID: \"5ae956fa-5a45-4248-b300-d802aa171d52\") " pod="metallb-system/speaker-pz9ts" Nov 24 17:15:19 crc kubenswrapper[4777]: I1124 17:15:19.092601 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-pz9ts" Nov 24 17:15:19 crc kubenswrapper[4777]: W1124 17:15:19.123764 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ae956fa_5a45_4248_b300_d802aa171d52.slice/crio-c09bc3af60f426b8f4382a2237d865555ae5b68771078514568de5c87bb1cdb8 WatchSource:0}: Error finding container c09bc3af60f426b8f4382a2237d865555ae5b68771078514568de5c87bb1cdb8: Status 404 returned error can't find the container with id c09bc3af60f426b8f4382a2237d865555ae5b68771078514568de5c87bb1cdb8 Nov 24 17:15:19 crc kubenswrapper[4777]: I1124 17:15:19.467759 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-pz9ts" event={"ID":"5ae956fa-5a45-4248-b300-d802aa171d52","Type":"ContainerStarted","Data":"5b85127cbd82c4a6caa768be70ffa01e6c9f3bac7dd68a2515a88c3016e2371d"} Nov 24 17:15:19 crc kubenswrapper[4777]: I1124 17:15:19.467800 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-pz9ts" event={"ID":"5ae956fa-5a45-4248-b300-d802aa171d52","Type":"ContainerStarted","Data":"c09bc3af60f426b8f4382a2237d865555ae5b68771078514568de5c87bb1cdb8"} Nov 24 17:15:19 crc kubenswrapper[4777]: I1124 17:15:19.471478 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-z9z24" event={"ID":"b4a78161-cf5f-4307-9adb-a05de51d620f","Type":"ContainerStarted","Data":"d2fed574f62ac115f5dfa26a50e9f66f8cac44f80391b291bc8c856c57f1ed74"} Nov 24 17:15:19 crc kubenswrapper[4777]: I1124 17:15:19.471510 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-z9z24" event={"ID":"b4a78161-cf5f-4307-9adb-a05de51d620f","Type":"ContainerStarted","Data":"1f4d8b2748b5764475f685ee21899db3a5cc497ea8704d357dd1f7fbf0561653"} Nov 24 17:15:19 crc kubenswrapper[4777]: I1124 17:15:19.471552 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-z9z24" event={"ID":"b4a78161-cf5f-4307-9adb-a05de51d620f","Type":"ContainerStarted","Data":"d6db0f499ee80b610aaef1d8993ebacab38486796839e969a514790fe0dea50a"} Nov 24 17:15:19 crc kubenswrapper[4777]: I1124 17:15:19.472345 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-z9z24" Nov 24 17:15:19 crc kubenswrapper[4777]: I1124 17:15:19.515839 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-z9z24" podStartSLOduration=2.515824014 podStartE2EDuration="2.515824014s" podCreationTimestamp="2025-11-24 17:15:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:15:19.514609579 +0000 UTC m=+877.673804628" watchObservedRunningTime="2025-11-24 17:15:19.515824014 +0000 UTC m=+877.675019053" Nov 24 17:15:20 crc kubenswrapper[4777]: I1124 17:15:20.486611 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-pz9ts" event={"ID":"5ae956fa-5a45-4248-b300-d802aa171d52","Type":"ContainerStarted","Data":"3ee91f6deeb48c92e3010bdcd60aa5da4f5a2c907e0809a46ea34e4b08b76617"} Nov 24 17:15:20 crc kubenswrapper[4777]: I1124 17:15:20.486905 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-pz9ts" Nov 24 17:15:20 crc kubenswrapper[4777]: I1124 17:15:20.502379 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-pz9ts" podStartSLOduration=3.502364296 podStartE2EDuration="3.502364296s" podCreationTimestamp="2025-11-24 17:15:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:15:20.50144233 +0000 UTC m=+878.660637369" watchObservedRunningTime="2025-11-24 17:15:20.502364296 +0000 UTC m=+878.661559345" Nov 24 17:15:25 crc kubenswrapper[4777]: I1124 17:15:25.528609 4777 generic.go:334] "Generic (PLEG): container finished" podID="0dd8e42b-e649-438d-bd92-e35a3ed06d4a" containerID="5a93097d909859891a3282a8521b69e2324a80bb4b71eeac84f354cb03e30fb1" exitCode=0 Nov 24 17:15:25 crc kubenswrapper[4777]: I1124 17:15:25.528654 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95bf" event={"ID":"0dd8e42b-e649-438d-bd92-e35a3ed06d4a","Type":"ContainerDied","Data":"5a93097d909859891a3282a8521b69e2324a80bb4b71eeac84f354cb03e30fb1"} Nov 24 17:15:25 crc kubenswrapper[4777]: I1124 17:15:25.533576 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-86vqd" event={"ID":"2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a","Type":"ContainerStarted","Data":"41bda508028af2f3dd712ec05d284109c67eff7843a4b1e48dd3ce2595272ba0"} Nov 24 17:15:25 crc kubenswrapper[4777]: I1124 17:15:25.533930 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-86vqd" Nov 24 17:15:25 crc kubenswrapper[4777]: I1124 17:15:25.586306 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-86vqd" podStartSLOduration=2.173103409 podStartE2EDuration="8.586282043s" podCreationTimestamp="2025-11-24 17:15:17 +0000 UTC" firstStartedPulling="2025-11-24 17:15:18.428860927 +0000 UTC m=+876.588055976" lastFinishedPulling="2025-11-24 17:15:24.842039561 +0000 UTC m=+883.001234610" observedRunningTime="2025-11-24 17:15:25.578474357 +0000 UTC m=+883.737669446" watchObservedRunningTime="2025-11-24 17:15:25.586282043 +0000 UTC m=+883.745477112" Nov 24 17:15:26 crc kubenswrapper[4777]: I1124 17:15:26.544513 4777 generic.go:334] "Generic (PLEG): container finished" podID="0dd8e42b-e649-438d-bd92-e35a3ed06d4a" containerID="500a49db2af7210235286aa830b31462c17d4b59051c8afdfaf07256fe103459" exitCode=0 Nov 24 17:15:26 crc kubenswrapper[4777]: I1124 17:15:26.545681 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95bf" event={"ID":"0dd8e42b-e649-438d-bd92-e35a3ed06d4a","Type":"ContainerDied","Data":"500a49db2af7210235286aa830b31462c17d4b59051c8afdfaf07256fe103459"} Nov 24 17:15:27 crc kubenswrapper[4777]: I1124 17:15:27.558585 4777 generic.go:334] "Generic (PLEG): container finished" podID="0dd8e42b-e649-438d-bd92-e35a3ed06d4a" containerID="9a17c3139a464a6c09d93350b328fc88127ec1584cfb4504b1b8760f04fcb158" exitCode=0 Nov 24 17:15:27 crc kubenswrapper[4777]: I1124 17:15:27.558731 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95bf" event={"ID":"0dd8e42b-e649-438d-bd92-e35a3ed06d4a","Type":"ContainerDied","Data":"9a17c3139a464a6c09d93350b328fc88127ec1584cfb4504b1b8760f04fcb158"} Nov 24 17:15:28 crc kubenswrapper[4777]: I1124 17:15:28.210828 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-z9z24" Nov 24 17:15:28 crc kubenswrapper[4777]: I1124 17:15:28.586084 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95bf" event={"ID":"0dd8e42b-e649-438d-bd92-e35a3ed06d4a","Type":"ContainerStarted","Data":"2236441632c02c85118679964509b969c4d13d345e8ff04f9ed459afd229794f"} Nov 24 17:15:28 crc kubenswrapper[4777]: I1124 17:15:28.586459 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95bf" event={"ID":"0dd8e42b-e649-438d-bd92-e35a3ed06d4a","Type":"ContainerStarted","Data":"5de86f6590de324974b0257060b133742d05c0b46ecd91dd63d12293d772660d"} Nov 24 17:15:28 crc kubenswrapper[4777]: I1124 17:15:28.586473 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95bf" event={"ID":"0dd8e42b-e649-438d-bd92-e35a3ed06d4a","Type":"ContainerStarted","Data":"15f084f291ea62a0d15d541650e8342481ea02ad8a0d46e689663661d4f36979"} Nov 24 17:15:28 crc kubenswrapper[4777]: I1124 17:15:28.586485 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95bf" event={"ID":"0dd8e42b-e649-438d-bd92-e35a3ed06d4a","Type":"ContainerStarted","Data":"81380cf6011144261810ed5f6946a4c92c865e0221c09aeea3ffa555ac41202b"} Nov 24 17:15:29 crc kubenswrapper[4777]: I1124 17:15:29.097240 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-pz9ts" Nov 24 17:15:29 crc kubenswrapper[4777]: I1124 17:15:29.597438 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95bf" event={"ID":"0dd8e42b-e649-438d-bd92-e35a3ed06d4a","Type":"ContainerStarted","Data":"1b4bc67eb194a3a4578f3da82b8bf68bd806d4c41889870a1e3c06dfab4692ec"} Nov 24 17:15:29 crc kubenswrapper[4777]: I1124 17:15:29.597492 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c95bf" event={"ID":"0dd8e42b-e649-438d-bd92-e35a3ed06d4a","Type":"ContainerStarted","Data":"d700c709564a47312f68393f9d01fdbfac1e95603890bd69cdf5c18721eb3e74"} Nov 24 17:15:29 crc kubenswrapper[4777]: I1124 17:15:29.597603 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:29 crc kubenswrapper[4777]: I1124 17:15:29.621461 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-c95bf" podStartSLOduration=5.473293449 podStartE2EDuration="12.621433807s" podCreationTimestamp="2025-11-24 17:15:17 +0000 UTC" firstStartedPulling="2025-11-24 17:15:17.655722608 +0000 UTC m=+875.814917657" lastFinishedPulling="2025-11-24 17:15:24.803862956 +0000 UTC m=+882.963058015" observedRunningTime="2025-11-24 17:15:29.619795229 +0000 UTC m=+887.778990329" watchObservedRunningTime="2025-11-24 17:15:29.621433807 +0000 UTC m=+887.780628896" Nov 24 17:15:32 crc kubenswrapper[4777]: I1124 17:15:32.067943 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-n8xgs"] Nov 24 17:15:32 crc kubenswrapper[4777]: I1124 17:15:32.069766 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n8xgs" Nov 24 17:15:32 crc kubenswrapper[4777]: I1124 17:15:32.097959 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 24 17:15:32 crc kubenswrapper[4777]: I1124 17:15:32.098011 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-4rjrq" Nov 24 17:15:32 crc kubenswrapper[4777]: I1124 17:15:32.098538 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 24 17:15:32 crc kubenswrapper[4777]: I1124 17:15:32.154204 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-n8xgs"] Nov 24 17:15:32 crc kubenswrapper[4777]: I1124 17:15:32.254071 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpsf5\" (UniqueName: \"kubernetes.io/projected/72d89195-06cc-4d27-937a-034a47f829e4-kube-api-access-kpsf5\") pod \"openstack-operator-index-n8xgs\" (UID: \"72d89195-06cc-4d27-937a-034a47f829e4\") " pod="openstack-operators/openstack-operator-index-n8xgs" Nov 24 17:15:32 crc kubenswrapper[4777]: I1124 17:15:32.355707 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpsf5\" (UniqueName: \"kubernetes.io/projected/72d89195-06cc-4d27-937a-034a47f829e4-kube-api-access-kpsf5\") pod \"openstack-operator-index-n8xgs\" (UID: \"72d89195-06cc-4d27-937a-034a47f829e4\") " pod="openstack-operators/openstack-operator-index-n8xgs" Nov 24 17:15:32 crc kubenswrapper[4777]: I1124 17:15:32.376476 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpsf5\" (UniqueName: \"kubernetes.io/projected/72d89195-06cc-4d27-937a-034a47f829e4-kube-api-access-kpsf5\") pod \"openstack-operator-index-n8xgs\" (UID: \"72d89195-06cc-4d27-937a-034a47f829e4\") " pod="openstack-operators/openstack-operator-index-n8xgs" Nov 24 17:15:32 crc kubenswrapper[4777]: I1124 17:15:32.414116 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n8xgs" Nov 24 17:15:32 crc kubenswrapper[4777]: I1124 17:15:32.507198 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:32 crc kubenswrapper[4777]: I1124 17:15:32.545489 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:32 crc kubenswrapper[4777]: I1124 17:15:32.854167 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-n8xgs"] Nov 24 17:15:33 crc kubenswrapper[4777]: I1124 17:15:33.659186 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n8xgs" event={"ID":"72d89195-06cc-4d27-937a-034a47f829e4","Type":"ContainerStarted","Data":"b6a9b836908ad9af10145a65dbcf48d811fb094de42c2d21e0c9c15bb7388b95"} Nov 24 17:15:35 crc kubenswrapper[4777]: I1124 17:15:35.424357 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-n8xgs"] Nov 24 17:15:36 crc kubenswrapper[4777]: I1124 17:15:36.056866 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zvxss"] Nov 24 17:15:36 crc kubenswrapper[4777]: I1124 17:15:36.057995 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zvxss" Nov 24 17:15:36 crc kubenswrapper[4777]: I1124 17:15:36.065462 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zvxss"] Nov 24 17:15:36 crc kubenswrapper[4777]: I1124 17:15:36.211294 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfs5w\" (UniqueName: \"kubernetes.io/projected/127dc4f2-329f-453e-8f21-1877225cec81-kube-api-access-qfs5w\") pod \"openstack-operator-index-zvxss\" (UID: \"127dc4f2-329f-453e-8f21-1877225cec81\") " pod="openstack-operators/openstack-operator-index-zvxss" Nov 24 17:15:36 crc kubenswrapper[4777]: I1124 17:15:36.313386 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfs5w\" (UniqueName: \"kubernetes.io/projected/127dc4f2-329f-453e-8f21-1877225cec81-kube-api-access-qfs5w\") pod \"openstack-operator-index-zvxss\" (UID: \"127dc4f2-329f-453e-8f21-1877225cec81\") " pod="openstack-operators/openstack-operator-index-zvxss" Nov 24 17:15:36 crc kubenswrapper[4777]: I1124 17:15:36.341213 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfs5w\" (UniqueName: \"kubernetes.io/projected/127dc4f2-329f-453e-8f21-1877225cec81-kube-api-access-qfs5w\") pod \"openstack-operator-index-zvxss\" (UID: \"127dc4f2-329f-453e-8f21-1877225cec81\") " pod="openstack-operators/openstack-operator-index-zvxss" Nov 24 17:15:36 crc kubenswrapper[4777]: I1124 17:15:36.392508 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zvxss" Nov 24 17:15:36 crc kubenswrapper[4777]: I1124 17:15:36.682313 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n8xgs" event={"ID":"72d89195-06cc-4d27-937a-034a47f829e4","Type":"ContainerStarted","Data":"40642e35626dd470477bf89740d07a4257d0172506ff29eb7c7dfb13326e0805"} Nov 24 17:15:36 crc kubenswrapper[4777]: I1124 17:15:36.682443 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-n8xgs" podUID="72d89195-06cc-4d27-937a-034a47f829e4" containerName="registry-server" containerID="cri-o://40642e35626dd470477bf89740d07a4257d0172506ff29eb7c7dfb13326e0805" gracePeriod=2 Nov 24 17:15:36 crc kubenswrapper[4777]: I1124 17:15:36.701685 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-n8xgs" podStartSLOduration=2.0423048599999998 podStartE2EDuration="4.70166822s" podCreationTimestamp="2025-11-24 17:15:32 +0000 UTC" firstStartedPulling="2025-11-24 17:15:32.867850501 +0000 UTC m=+891.027045590" lastFinishedPulling="2025-11-24 17:15:35.527213891 +0000 UTC m=+893.686408950" observedRunningTime="2025-11-24 17:15:36.698309923 +0000 UTC m=+894.857504972" watchObservedRunningTime="2025-11-24 17:15:36.70166822 +0000 UTC m=+894.860863269" Nov 24 17:15:36 crc kubenswrapper[4777]: I1124 17:15:36.871037 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zvxss"] Nov 24 17:15:36 crc kubenswrapper[4777]: W1124 17:15:36.873478 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod127dc4f2_329f_453e_8f21_1877225cec81.slice/crio-5e72c8533fb416f315e0999b0a8ae074b556fb862c75578120fee38e1cb04878 WatchSource:0}: Error finding container 5e72c8533fb416f315e0999b0a8ae074b556fb862c75578120fee38e1cb04878: Status 404 returned error can't find the container with id 5e72c8533fb416f315e0999b0a8ae074b556fb862c75578120fee38e1cb04878 Nov 24 17:15:36 crc kubenswrapper[4777]: I1124 17:15:36.878988 4777 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.134027 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n8xgs" Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.225898 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpsf5\" (UniqueName: \"kubernetes.io/projected/72d89195-06cc-4d27-937a-034a47f829e4-kube-api-access-kpsf5\") pod \"72d89195-06cc-4d27-937a-034a47f829e4\" (UID: \"72d89195-06cc-4d27-937a-034a47f829e4\") " Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.231512 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72d89195-06cc-4d27-937a-034a47f829e4-kube-api-access-kpsf5" (OuterVolumeSpecName: "kube-api-access-kpsf5") pod "72d89195-06cc-4d27-937a-034a47f829e4" (UID: "72d89195-06cc-4d27-937a-034a47f829e4"). InnerVolumeSpecName "kube-api-access-kpsf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.328893 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpsf5\" (UniqueName: \"kubernetes.io/projected/72d89195-06cc-4d27-937a-034a47f829e4-kube-api-access-kpsf5\") on node \"crc\" DevicePath \"\"" Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.511570 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-c95bf" Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.693213 4777 generic.go:334] "Generic (PLEG): container finished" podID="72d89195-06cc-4d27-937a-034a47f829e4" containerID="40642e35626dd470477bf89740d07a4257d0172506ff29eb7c7dfb13326e0805" exitCode=0 Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.693293 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n8xgs" Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.693346 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n8xgs" event={"ID":"72d89195-06cc-4d27-937a-034a47f829e4","Type":"ContainerDied","Data":"40642e35626dd470477bf89740d07a4257d0172506ff29eb7c7dfb13326e0805"} Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.693421 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n8xgs" event={"ID":"72d89195-06cc-4d27-937a-034a47f829e4","Type":"ContainerDied","Data":"b6a9b836908ad9af10145a65dbcf48d811fb094de42c2d21e0c9c15bb7388b95"} Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.693459 4777 scope.go:117] "RemoveContainer" containerID="40642e35626dd470477bf89740d07a4257d0172506ff29eb7c7dfb13326e0805" Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.697121 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zvxss" event={"ID":"127dc4f2-329f-453e-8f21-1877225cec81","Type":"ContainerStarted","Data":"4c72f6e0cc93e5f47c827b7426d5b79f7b21f84250f54a4742ab4e6df75cc894"} Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.697195 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zvxss" event={"ID":"127dc4f2-329f-453e-8f21-1877225cec81","Type":"ContainerStarted","Data":"5e72c8533fb416f315e0999b0a8ae074b556fb862c75578120fee38e1cb04878"} Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.721051 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-n8xgs"] Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.721563 4777 scope.go:117] "RemoveContainer" containerID="40642e35626dd470477bf89740d07a4257d0172506ff29eb7c7dfb13326e0805" Nov 24 17:15:37 crc kubenswrapper[4777]: E1124 17:15:37.722238 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40642e35626dd470477bf89740d07a4257d0172506ff29eb7c7dfb13326e0805\": container with ID starting with 40642e35626dd470477bf89740d07a4257d0172506ff29eb7c7dfb13326e0805 not found: ID does not exist" containerID="40642e35626dd470477bf89740d07a4257d0172506ff29eb7c7dfb13326e0805" Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.722291 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40642e35626dd470477bf89740d07a4257d0172506ff29eb7c7dfb13326e0805"} err="failed to get container status \"40642e35626dd470477bf89740d07a4257d0172506ff29eb7c7dfb13326e0805\": rpc error: code = NotFound desc = could not find container \"40642e35626dd470477bf89740d07a4257d0172506ff29eb7c7dfb13326e0805\": container with ID starting with 40642e35626dd470477bf89740d07a4257d0172506ff29eb7c7dfb13326e0805 not found: ID does not exist" Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.733889 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-n8xgs"] Nov 24 17:15:37 crc kubenswrapper[4777]: I1124 17:15:37.737721 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zvxss" podStartSLOduration=1.665546316 podStartE2EDuration="1.737692433s" podCreationTimestamp="2025-11-24 17:15:36 +0000 UTC" firstStartedPulling="2025-11-24 17:15:36.87865378 +0000 UTC m=+895.037848829" lastFinishedPulling="2025-11-24 17:15:36.950799857 +0000 UTC m=+895.109994946" observedRunningTime="2025-11-24 17:15:37.733642966 +0000 UTC m=+895.892838075" watchObservedRunningTime="2025-11-24 17:15:37.737692433 +0000 UTC m=+895.896887522" Nov 24 17:15:38 crc kubenswrapper[4777]: I1124 17:15:38.124124 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-86vqd" Nov 24 17:15:39 crc kubenswrapper[4777]: I1124 17:15:39.261532 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72d89195-06cc-4d27-937a-034a47f829e4" path="/var/lib/kubelet/pods/72d89195-06cc-4d27-937a-034a47f829e4/volumes" Nov 24 17:15:44 crc kubenswrapper[4777]: I1124 17:15:44.270568 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:15:44 crc kubenswrapper[4777]: I1124 17:15:44.271064 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:15:46 crc kubenswrapper[4777]: I1124 17:15:46.393260 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-zvxss" Nov 24 17:15:46 crc kubenswrapper[4777]: I1124 17:15:46.394094 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-zvxss" Nov 24 17:15:46 crc kubenswrapper[4777]: I1124 17:15:46.435770 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-zvxss" Nov 24 17:15:46 crc kubenswrapper[4777]: I1124 17:15:46.825357 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-zvxss" Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.291947 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl"] Nov 24 17:15:48 crc kubenswrapper[4777]: E1124 17:15:48.292746 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d89195-06cc-4d27-937a-034a47f829e4" containerName="registry-server" Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.292769 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d89195-06cc-4d27-937a-034a47f829e4" containerName="registry-server" Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.293017 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="72d89195-06cc-4d27-937a-034a47f829e4" containerName="registry-server" Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.294570 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.297647 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-5z74h" Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.317062 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl"] Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.402544 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-bundle\") pod \"487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl\" (UID: \"bd350075-1ebb-4431-9b02-91e2d6c5e7d3\") " pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.402689 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqdn4\" (UniqueName: \"kubernetes.io/projected/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-kube-api-access-pqdn4\") pod \"487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl\" (UID: \"bd350075-1ebb-4431-9b02-91e2d6c5e7d3\") " pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.402897 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-util\") pod \"487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl\" (UID: \"bd350075-1ebb-4431-9b02-91e2d6c5e7d3\") " pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.504415 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-bundle\") pod \"487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl\" (UID: \"bd350075-1ebb-4431-9b02-91e2d6c5e7d3\") " pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.504530 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqdn4\" (UniqueName: \"kubernetes.io/projected/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-kube-api-access-pqdn4\") pod \"487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl\" (UID: \"bd350075-1ebb-4431-9b02-91e2d6c5e7d3\") " pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.504666 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-util\") pod \"487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl\" (UID: \"bd350075-1ebb-4431-9b02-91e2d6c5e7d3\") " pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.505231 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-bundle\") pod \"487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl\" (UID: \"bd350075-1ebb-4431-9b02-91e2d6c5e7d3\") " pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.505401 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-util\") pod \"487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl\" (UID: \"bd350075-1ebb-4431-9b02-91e2d6c5e7d3\") " pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.543343 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqdn4\" (UniqueName: \"kubernetes.io/projected/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-kube-api-access-pqdn4\") pod \"487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl\" (UID: \"bd350075-1ebb-4431-9b02-91e2d6c5e7d3\") " pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" Nov 24 17:15:48 crc kubenswrapper[4777]: I1124 17:15:48.616489 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" Nov 24 17:15:49 crc kubenswrapper[4777]: I1124 17:15:49.115016 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl"] Nov 24 17:15:49 crc kubenswrapper[4777]: W1124 17:15:49.120495 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd350075_1ebb_4431_9b02_91e2d6c5e7d3.slice/crio-1bf78fb88b44f70257f6e51cddfe2278db7b70450e31a4c8d4f4fd0270831261 WatchSource:0}: Error finding container 1bf78fb88b44f70257f6e51cddfe2278db7b70450e31a4c8d4f4fd0270831261: Status 404 returned error can't find the container with id 1bf78fb88b44f70257f6e51cddfe2278db7b70450e31a4c8d4f4fd0270831261 Nov 24 17:15:49 crc kubenswrapper[4777]: I1124 17:15:49.809177 4777 generic.go:334] "Generic (PLEG): container finished" podID="bd350075-1ebb-4431-9b02-91e2d6c5e7d3" containerID="ed652dce55ac8aa1b25ae156bd3582eed7cd69f7c288abe1ebbbc09bad3d3da3" exitCode=0 Nov 24 17:15:49 crc kubenswrapper[4777]: I1124 17:15:49.809259 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" event={"ID":"bd350075-1ebb-4431-9b02-91e2d6c5e7d3","Type":"ContainerDied","Data":"ed652dce55ac8aa1b25ae156bd3582eed7cd69f7c288abe1ebbbc09bad3d3da3"} Nov 24 17:15:49 crc kubenswrapper[4777]: I1124 17:15:49.810758 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" event={"ID":"bd350075-1ebb-4431-9b02-91e2d6c5e7d3","Type":"ContainerStarted","Data":"1bf78fb88b44f70257f6e51cddfe2278db7b70450e31a4c8d4f4fd0270831261"} Nov 24 17:15:50 crc kubenswrapper[4777]: I1124 17:15:50.825397 4777 generic.go:334] "Generic (PLEG): container finished" podID="bd350075-1ebb-4431-9b02-91e2d6c5e7d3" containerID="e01a946e2b833e890ba96cfd04ca2bccf9429ddf0593b089a4d602759cd063ea" exitCode=0 Nov 24 17:15:50 crc kubenswrapper[4777]: I1124 17:15:50.825518 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" event={"ID":"bd350075-1ebb-4431-9b02-91e2d6c5e7d3","Type":"ContainerDied","Data":"e01a946e2b833e890ba96cfd04ca2bccf9429ddf0593b089a4d602759cd063ea"} Nov 24 17:15:51 crc kubenswrapper[4777]: I1124 17:15:51.847053 4777 generic.go:334] "Generic (PLEG): container finished" podID="bd350075-1ebb-4431-9b02-91e2d6c5e7d3" containerID="fb5e11977293b50ac6c5349da2488b924bd690214f583ddfd41bf21891ab7f74" exitCode=0 Nov 24 17:15:51 crc kubenswrapper[4777]: I1124 17:15:51.847124 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" event={"ID":"bd350075-1ebb-4431-9b02-91e2d6c5e7d3","Type":"ContainerDied","Data":"fb5e11977293b50ac6c5349da2488b924bd690214f583ddfd41bf21891ab7f74"} Nov 24 17:15:53 crc kubenswrapper[4777]: I1124 17:15:53.229018 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" Nov 24 17:15:53 crc kubenswrapper[4777]: I1124 17:15:53.381738 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-bundle\") pod \"bd350075-1ebb-4431-9b02-91e2d6c5e7d3\" (UID: \"bd350075-1ebb-4431-9b02-91e2d6c5e7d3\") " Nov 24 17:15:53 crc kubenswrapper[4777]: I1124 17:15:53.382025 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-util\") pod \"bd350075-1ebb-4431-9b02-91e2d6c5e7d3\" (UID: \"bd350075-1ebb-4431-9b02-91e2d6c5e7d3\") " Nov 24 17:15:53 crc kubenswrapper[4777]: I1124 17:15:53.382252 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqdn4\" (UniqueName: \"kubernetes.io/projected/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-kube-api-access-pqdn4\") pod \"bd350075-1ebb-4431-9b02-91e2d6c5e7d3\" (UID: \"bd350075-1ebb-4431-9b02-91e2d6c5e7d3\") " Nov 24 17:15:53 crc kubenswrapper[4777]: I1124 17:15:53.384352 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-bundle" (OuterVolumeSpecName: "bundle") pod "bd350075-1ebb-4431-9b02-91e2d6c5e7d3" (UID: "bd350075-1ebb-4431-9b02-91e2d6c5e7d3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:15:53 crc kubenswrapper[4777]: I1124 17:15:53.394556 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-kube-api-access-pqdn4" (OuterVolumeSpecName: "kube-api-access-pqdn4") pod "bd350075-1ebb-4431-9b02-91e2d6c5e7d3" (UID: "bd350075-1ebb-4431-9b02-91e2d6c5e7d3"). InnerVolumeSpecName "kube-api-access-pqdn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:15:53 crc kubenswrapper[4777]: I1124 17:15:53.400939 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-util" (OuterVolumeSpecName: "util") pod "bd350075-1ebb-4431-9b02-91e2d6c5e7d3" (UID: "bd350075-1ebb-4431-9b02-91e2d6c5e7d3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:15:53 crc kubenswrapper[4777]: I1124 17:15:53.485392 4777 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-util\") on node \"crc\" DevicePath \"\"" Nov 24 17:15:53 crc kubenswrapper[4777]: I1124 17:15:53.485440 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqdn4\" (UniqueName: \"kubernetes.io/projected/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-kube-api-access-pqdn4\") on node \"crc\" DevicePath \"\"" Nov 24 17:15:53 crc kubenswrapper[4777]: I1124 17:15:53.485465 4777 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd350075-1ebb-4431-9b02-91e2d6c5e7d3-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:15:53 crc kubenswrapper[4777]: I1124 17:15:53.867852 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" event={"ID":"bd350075-1ebb-4431-9b02-91e2d6c5e7d3","Type":"ContainerDied","Data":"1bf78fb88b44f70257f6e51cddfe2278db7b70450e31a4c8d4f4fd0270831261"} Nov 24 17:15:53 crc kubenswrapper[4777]: I1124 17:15:53.868293 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bf78fb88b44f70257f6e51cddfe2278db7b70450e31a4c8d4f4fd0270831261" Nov 24 17:15:53 crc kubenswrapper[4777]: I1124 17:15:53.868013 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl" Nov 24 17:16:01 crc kubenswrapper[4777]: I1124 17:16:01.799482 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-68bb788c6f-ft4rz"] Nov 24 17:16:01 crc kubenswrapper[4777]: E1124 17:16:01.801622 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd350075-1ebb-4431-9b02-91e2d6c5e7d3" containerName="util" Nov 24 17:16:01 crc kubenswrapper[4777]: I1124 17:16:01.801745 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd350075-1ebb-4431-9b02-91e2d6c5e7d3" containerName="util" Nov 24 17:16:01 crc kubenswrapper[4777]: E1124 17:16:01.801845 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd350075-1ebb-4431-9b02-91e2d6c5e7d3" containerName="pull" Nov 24 17:16:01 crc kubenswrapper[4777]: I1124 17:16:01.801926 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd350075-1ebb-4431-9b02-91e2d6c5e7d3" containerName="pull" Nov 24 17:16:01 crc kubenswrapper[4777]: E1124 17:16:01.802151 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd350075-1ebb-4431-9b02-91e2d6c5e7d3" containerName="extract" Nov 24 17:16:01 crc kubenswrapper[4777]: I1124 17:16:01.802235 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd350075-1ebb-4431-9b02-91e2d6c5e7d3" containerName="extract" Nov 24 17:16:01 crc kubenswrapper[4777]: I1124 17:16:01.802675 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd350075-1ebb-4431-9b02-91e2d6c5e7d3" containerName="extract" Nov 24 17:16:01 crc kubenswrapper[4777]: I1124 17:16:01.804262 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-68bb788c6f-ft4rz" Nov 24 17:16:01 crc kubenswrapper[4777]: I1124 17:16:01.812559 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-4zdjc" Nov 24 17:16:01 crc kubenswrapper[4777]: I1124 17:16:01.833268 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-68bb788c6f-ft4rz"] Nov 24 17:16:01 crc kubenswrapper[4777]: I1124 17:16:01.926194 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqfhm\" (UniqueName: \"kubernetes.io/projected/547759be-2d04-4553-aedc-e3f1b7ef9779-kube-api-access-xqfhm\") pod \"openstack-operator-controller-operator-68bb788c6f-ft4rz\" (UID: \"547759be-2d04-4553-aedc-e3f1b7ef9779\") " pod="openstack-operators/openstack-operator-controller-operator-68bb788c6f-ft4rz" Nov 24 17:16:02 crc kubenswrapper[4777]: I1124 17:16:02.027312 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqfhm\" (UniqueName: \"kubernetes.io/projected/547759be-2d04-4553-aedc-e3f1b7ef9779-kube-api-access-xqfhm\") pod \"openstack-operator-controller-operator-68bb788c6f-ft4rz\" (UID: \"547759be-2d04-4553-aedc-e3f1b7ef9779\") " pod="openstack-operators/openstack-operator-controller-operator-68bb788c6f-ft4rz" Nov 24 17:16:02 crc kubenswrapper[4777]: I1124 17:16:02.044351 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqfhm\" (UniqueName: \"kubernetes.io/projected/547759be-2d04-4553-aedc-e3f1b7ef9779-kube-api-access-xqfhm\") pod \"openstack-operator-controller-operator-68bb788c6f-ft4rz\" (UID: \"547759be-2d04-4553-aedc-e3f1b7ef9779\") " pod="openstack-operators/openstack-operator-controller-operator-68bb788c6f-ft4rz" Nov 24 17:16:02 crc kubenswrapper[4777]: I1124 17:16:02.137994 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-68bb788c6f-ft4rz" Nov 24 17:16:02 crc kubenswrapper[4777]: I1124 17:16:02.605862 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-68bb788c6f-ft4rz"] Nov 24 17:16:02 crc kubenswrapper[4777]: I1124 17:16:02.940494 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-68bb788c6f-ft4rz" event={"ID":"547759be-2d04-4553-aedc-e3f1b7ef9779","Type":"ContainerStarted","Data":"4a34dde477d01b91c9e4b338525c45d80d61bcc4906cc93dacbee6bfdcd5e69b"} Nov 24 17:16:06 crc kubenswrapper[4777]: I1124 17:16:06.974741 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-68bb788c6f-ft4rz" event={"ID":"547759be-2d04-4553-aedc-e3f1b7ef9779","Type":"ContainerStarted","Data":"e6d05cf2a54c700d3f00720fdf49be06a05a8953a916323a67a71e2ae1ed7ea1"} Nov 24 17:16:08 crc kubenswrapper[4777]: I1124 17:16:08.990411 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-68bb788c6f-ft4rz" event={"ID":"547759be-2d04-4553-aedc-e3f1b7ef9779","Type":"ContainerStarted","Data":"7673ef9d29ab5631d3979d9d81a74493bf45e885b5d0d407ba1661dcc963f034"} Nov 24 17:16:08 crc kubenswrapper[4777]: I1124 17:16:08.990793 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-68bb788c6f-ft4rz" Nov 24 17:16:09 crc kubenswrapper[4777]: I1124 17:16:09.018214 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-68bb788c6f-ft4rz" podStartSLOduration=2.217536627 podStartE2EDuration="8.018195081s" podCreationTimestamp="2025-11-24 17:16:01 +0000 UTC" firstStartedPulling="2025-11-24 17:16:02.610927057 +0000 UTC m=+920.770122106" lastFinishedPulling="2025-11-24 17:16:08.411585511 +0000 UTC m=+926.570780560" observedRunningTime="2025-11-24 17:16:09.016692018 +0000 UTC m=+927.175887067" watchObservedRunningTime="2025-11-24 17:16:09.018195081 +0000 UTC m=+927.177390140" Nov 24 17:16:12 crc kubenswrapper[4777]: I1124 17:16:12.141122 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-68bb788c6f-ft4rz" Nov 24 17:16:14 crc kubenswrapper[4777]: I1124 17:16:14.270640 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:16:14 crc kubenswrapper[4777]: I1124 17:16:14.270739 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:16:14 crc kubenswrapper[4777]: I1124 17:16:14.270810 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:16:14 crc kubenswrapper[4777]: I1124 17:16:14.271828 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"90d22c4d0e8b2ba20584ba22b6a107d623c9ded1ab193fb10e43b251d31391cc"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:16:14 crc kubenswrapper[4777]: I1124 17:16:14.271948 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://90d22c4d0e8b2ba20584ba22b6a107d623c9ded1ab193fb10e43b251d31391cc" gracePeriod=600 Nov 24 17:16:15 crc kubenswrapper[4777]: I1124 17:16:15.039077 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="90d22c4d0e8b2ba20584ba22b6a107d623c9ded1ab193fb10e43b251d31391cc" exitCode=0 Nov 24 17:16:15 crc kubenswrapper[4777]: I1124 17:16:15.039350 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"90d22c4d0e8b2ba20584ba22b6a107d623c9ded1ab193fb10e43b251d31391cc"} Nov 24 17:16:15 crc kubenswrapper[4777]: I1124 17:16:15.039584 4777 scope.go:117] "RemoveContainer" containerID="a0f6b6c1a5fc2ed6243c1d079f9f074d4a99e988147412f7b07f358cd4f1c26d" Nov 24 17:16:16 crc kubenswrapper[4777]: I1124 17:16:16.051599 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"3cf02e0241e501ffd35b75c9b5e93243f3fad1bfd50b46e463488c0fec7f7c53"} Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.056358 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-2chns"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.058796 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-2chns" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.060652 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-fqg2q" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.061148 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-82ppc"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.062532 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-82ppc" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.072450 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-8mj7q" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.077315 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-2chns"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.082738 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-82ppc"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.091806 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-klpxr"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.093159 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-klpxr" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.098472 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-9gkmd" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.103136 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-klpxr"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.116519 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-9mzvw"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.117420 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-9mzvw" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.129179 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-8gx8g" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.156612 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-zmms9"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.157660 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-zmms9" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.161561 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-kgrl8" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.182168 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-9mzvw"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.195507 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-zmms9"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.198496 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpb9b\" (UniqueName: \"kubernetes.io/projected/99bfc986-b80d-4193-bf32-4488c11f0066-kube-api-access-gpb9b\") pod \"designate-operator-controller-manager-767ccfd65f-klpxr\" (UID: \"99bfc986-b80d-4193-bf32-4488c11f0066\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-klpxr" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.198547 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvzsg\" (UniqueName: \"kubernetes.io/projected/1d3fc8a3-f3cf-4f37-846d-b244a52415f8-kube-api-access-pvzsg\") pod \"heat-operator-controller-manager-56f54d6746-zmms9\" (UID: \"1d3fc8a3-f3cf-4f37-846d-b244a52415f8\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-zmms9" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.198591 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs9vj\" (UniqueName: \"kubernetes.io/projected/a7457835-23e4-460e-a155-07c1d4e7c30e-kube-api-access-qs9vj\") pod \"barbican-operator-controller-manager-75fb479bcc-82ppc\" (UID: \"a7457835-23e4-460e-a155-07c1d4e7c30e\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-82ppc" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.198632 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dzjk\" (UniqueName: \"kubernetes.io/projected/fe817743-7f1e-46a4-9757-5b678ddc097a-kube-api-access-5dzjk\") pod \"glance-operator-controller-manager-7969689c84-9mzvw\" (UID: \"fe817743-7f1e-46a4-9757-5b678ddc097a\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-9mzvw" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.198670 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m98s\" (UniqueName: \"kubernetes.io/projected/1d2ecc53-902f-41f3-80f4-93f9e48c8532-kube-api-access-2m98s\") pod \"cinder-operator-controller-manager-6498cbf48f-2chns\" (UID: \"1d2ecc53-902f-41f3-80f4-93f9e48c8532\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-2chns" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.225687 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-6xxk9"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.227024 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-6xxk9" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.233117 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.233287 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-7kzrc" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.234447 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.238817 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-wmvls"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.239852 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-wmvls" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.240407 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-69q9n" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.242716 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.251333 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-ff2qj" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.251788 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.253006 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.263731 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-bl97z" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.272275 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-wmvls"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.299829 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dzjk\" (UniqueName: \"kubernetes.io/projected/fe817743-7f1e-46a4-9757-5b678ddc097a-kube-api-access-5dzjk\") pod \"glance-operator-controller-manager-7969689c84-9mzvw\" (UID: \"fe817743-7f1e-46a4-9757-5b678ddc097a\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-9mzvw" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.299889 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m98s\" (UniqueName: \"kubernetes.io/projected/1d2ecc53-902f-41f3-80f4-93f9e48c8532-kube-api-access-2m98s\") pod \"cinder-operator-controller-manager-6498cbf48f-2chns\" (UID: \"1d2ecc53-902f-41f3-80f4-93f9e48c8532\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-2chns" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.299922 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf68z\" (UniqueName: \"kubernetes.io/projected/8401afce-ddb4-4195-b6ae-7e5c91128525-kube-api-access-tf68z\") pod \"horizon-operator-controller-manager-598f69df5d-6xxk9\" (UID: \"8401afce-ddb4-4195-b6ae-7e5c91128525\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-6xxk9" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.299980 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpb9b\" (UniqueName: \"kubernetes.io/projected/99bfc986-b80d-4193-bf32-4488c11f0066-kube-api-access-gpb9b\") pod \"designate-operator-controller-manager-767ccfd65f-klpxr\" (UID: \"99bfc986-b80d-4193-bf32-4488c11f0066\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-klpxr" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.300015 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvzsg\" (UniqueName: \"kubernetes.io/projected/1d3fc8a3-f3cf-4f37-846d-b244a52415f8-kube-api-access-pvzsg\") pod \"heat-operator-controller-manager-56f54d6746-zmms9\" (UID: \"1d3fc8a3-f3cf-4f37-846d-b244a52415f8\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-zmms9" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.300034 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qs9vj\" (UniqueName: \"kubernetes.io/projected/a7457835-23e4-460e-a155-07c1d4e7c30e-kube-api-access-qs9vj\") pod \"barbican-operator-controller-manager-75fb479bcc-82ppc\" (UID: \"a7457835-23e4-460e-a155-07c1d4e7c30e\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-82ppc" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.302297 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.314131 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-6xxk9"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.334581 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qs9vj\" (UniqueName: \"kubernetes.io/projected/a7457835-23e4-460e-a155-07c1d4e7c30e-kube-api-access-qs9vj\") pod \"barbican-operator-controller-manager-75fb479bcc-82ppc\" (UID: \"a7457835-23e4-460e-a155-07c1d4e7c30e\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-82ppc" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.343614 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dzjk\" (UniqueName: \"kubernetes.io/projected/fe817743-7f1e-46a4-9757-5b678ddc097a-kube-api-access-5dzjk\") pod \"glance-operator-controller-manager-7969689c84-9mzvw\" (UID: \"fe817743-7f1e-46a4-9757-5b678ddc097a\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-9mzvw" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.348527 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvzsg\" (UniqueName: \"kubernetes.io/projected/1d3fc8a3-f3cf-4f37-846d-b244a52415f8-kube-api-access-pvzsg\") pod \"heat-operator-controller-manager-56f54d6746-zmms9\" (UID: \"1d3fc8a3-f3cf-4f37-846d-b244a52415f8\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-zmms9" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.375098 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.382952 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpb9b\" (UniqueName: \"kubernetes.io/projected/99bfc986-b80d-4193-bf32-4488c11f0066-kube-api-access-gpb9b\") pod \"designate-operator-controller-manager-767ccfd65f-klpxr\" (UID: \"99bfc986-b80d-4193-bf32-4488c11f0066\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-klpxr" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.384538 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m98s\" (UniqueName: \"kubernetes.io/projected/1d2ecc53-902f-41f3-80f4-93f9e48c8532-kube-api-access-2m98s\") pod \"cinder-operator-controller-manager-6498cbf48f-2chns\" (UID: \"1d2ecc53-902f-41f3-80f4-93f9e48c8532\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-2chns" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.395379 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-82ppc" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.401738 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf68z\" (UniqueName: \"kubernetes.io/projected/8401afce-ddb4-4195-b6ae-7e5c91128525-kube-api-access-tf68z\") pod \"horizon-operator-controller-manager-598f69df5d-6xxk9\" (UID: \"8401afce-ddb4-4195-b6ae-7e5c91128525\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-6xxk9" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.402866 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b755cc8b-dd00-4b28-8fb4-908e0e3db9a5-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-r5pxc\" (UID: \"b755cc8b-dd00-4b28-8fb4-908e0e3db9a5\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.402951 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97vbm\" (UniqueName: \"kubernetes.io/projected/b755cc8b-dd00-4b28-8fb4-908e0e3db9a5-kube-api-access-97vbm\") pod \"infra-operator-controller-manager-6dd8864d7c-r5pxc\" (UID: \"b755cc8b-dd00-4b28-8fb4-908e0e3db9a5\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.403056 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq2cw\" (UniqueName: \"kubernetes.io/projected/f7f44c61-027a-41c4-9665-8394d579ba33-kube-api-access-rq2cw\") pod \"ironic-operator-controller-manager-99b499f4-wmvls\" (UID: \"f7f44c61-027a-41c4-9665-8394d579ba33\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-wmvls" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.403136 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrxv7\" (UniqueName: \"kubernetes.io/projected/f1425713-2fd6-4108-abe3-9f46063e98f5-kube-api-access-vrxv7\") pod \"keystone-operator-controller-manager-7454b96578-v5vgx\" (UID: \"f1425713-2fd6-4108-abe3-9f46063e98f5\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.416216 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-klpxr" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.416592 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-g9kpn"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.417535 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-g9kpn" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.430155 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-dc68g"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.431148 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-dc68g" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.431644 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-9mzvw" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.432127 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-7gwf9" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.447332 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-wn8zd" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.449753 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.450835 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf68z\" (UniqueName: \"kubernetes.io/projected/8401afce-ddb4-4195-b6ae-7e5c91128525-kube-api-access-tf68z\") pod \"horizon-operator-controller-manager-598f69df5d-6xxk9\" (UID: \"8401afce-ddb4-4195-b6ae-7e5c91128525\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-6xxk9" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.451212 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.456511 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-kxtqm" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.488961 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-zmms9" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.507988 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-dc68g"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.509894 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b755cc8b-dd00-4b28-8fb4-908e0e3db9a5-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-r5pxc\" (UID: \"b755cc8b-dd00-4b28-8fb4-908e0e3db9a5\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.509928 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvn88\" (UniqueName: \"kubernetes.io/projected/fd1db7a5-22e3-4f16-a026-0b8895ef292d-kube-api-access-cvn88\") pod \"neutron-operator-controller-manager-78bd47f458-dc68g\" (UID: \"fd1db7a5-22e3-4f16-a026-0b8895ef292d\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-dc68g" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.509952 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97vbm\" (UniqueName: \"kubernetes.io/projected/b755cc8b-dd00-4b28-8fb4-908e0e3db9a5-kube-api-access-97vbm\") pod \"infra-operator-controller-manager-6dd8864d7c-r5pxc\" (UID: \"b755cc8b-dd00-4b28-8fb4-908e0e3db9a5\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.509996 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq2cw\" (UniqueName: \"kubernetes.io/projected/f7f44c61-027a-41c4-9665-8394d579ba33-kube-api-access-rq2cw\") pod \"ironic-operator-controller-manager-99b499f4-wmvls\" (UID: \"f7f44c61-027a-41c4-9665-8394d579ba33\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-wmvls" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.510013 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrxv7\" (UniqueName: \"kubernetes.io/projected/f1425713-2fd6-4108-abe3-9f46063e98f5-kube-api-access-vrxv7\") pod \"keystone-operator-controller-manager-7454b96578-v5vgx\" (UID: \"f1425713-2fd6-4108-abe3-9f46063e98f5\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.510041 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cpv8\" (UniqueName: \"kubernetes.io/projected/73a88858-0760-4b40-b57f-71e4f9977129-kube-api-access-2cpv8\") pod \"mariadb-operator-controller-manager-54b5986bb8-g9kpn\" (UID: \"73a88858-0760-4b40-b57f-71e4f9977129\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-g9kpn" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.510056 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv77l\" (UniqueName: \"kubernetes.io/projected/38f287c7-6c43-4dc9-b547-0d494704874a-kube-api-access-fv77l\") pod \"manila-operator-controller-manager-58f887965d-mf8r8\" (UID: \"38f287c7-6c43-4dc9-b547-0d494704874a\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.537618 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b755cc8b-dd00-4b28-8fb4-908e0e3db9a5-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-r5pxc\" (UID: \"b755cc8b-dd00-4b28-8fb4-908e0e3db9a5\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.546252 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-g9kpn"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.546321 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.548296 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-6xxk9" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.564033 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-ss948"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.564426 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq2cw\" (UniqueName: \"kubernetes.io/projected/f7f44c61-027a-41c4-9665-8394d579ba33-kube-api-access-rq2cw\") pod \"ironic-operator-controller-manager-99b499f4-wmvls\" (UID: \"f7f44c61-027a-41c4-9665-8394d579ba33\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-wmvls" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.598796 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrxv7\" (UniqueName: \"kubernetes.io/projected/f1425713-2fd6-4108-abe3-9f46063e98f5-kube-api-access-vrxv7\") pod \"keystone-operator-controller-manager-7454b96578-v5vgx\" (UID: \"f1425713-2fd6-4108-abe3-9f46063e98f5\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.602353 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-ss948" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.604096 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.606607 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-wmvls" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.609936 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-8dv9z" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.620046 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvn88\" (UniqueName: \"kubernetes.io/projected/fd1db7a5-22e3-4f16-a026-0b8895ef292d-kube-api-access-cvn88\") pod \"neutron-operator-controller-manager-78bd47f458-dc68g\" (UID: \"fd1db7a5-22e3-4f16-a026-0b8895ef292d\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-dc68g" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.620137 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cpv8\" (UniqueName: \"kubernetes.io/projected/73a88858-0760-4b40-b57f-71e4f9977129-kube-api-access-2cpv8\") pod \"mariadb-operator-controller-manager-54b5986bb8-g9kpn\" (UID: \"73a88858-0760-4b40-b57f-71e4f9977129\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-g9kpn" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.620163 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv77l\" (UniqueName: \"kubernetes.io/projected/38f287c7-6c43-4dc9-b547-0d494704874a-kube-api-access-fv77l\") pod \"manila-operator-controller-manager-58f887965d-mf8r8\" (UID: \"38f287c7-6c43-4dc9-b547-0d494704874a\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.651209 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97vbm\" (UniqueName: \"kubernetes.io/projected/b755cc8b-dd00-4b28-8fb4-908e0e3db9a5-kube-api-access-97vbm\") pod \"infra-operator-controller-manager-6dd8864d7c-r5pxc\" (UID: \"b755cc8b-dd00-4b28-8fb4-908e0e3db9a5\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.652843 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-ss948"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.670636 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvn88\" (UniqueName: \"kubernetes.io/projected/fd1db7a5-22e3-4f16-a026-0b8895ef292d-kube-api-access-cvn88\") pod \"neutron-operator-controller-manager-78bd47f458-dc68g\" (UID: \"fd1db7a5-22e3-4f16-a026-0b8895ef292d\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-dc68g" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.682848 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cpv8\" (UniqueName: \"kubernetes.io/projected/73a88858-0760-4b40-b57f-71e4f9977129-kube-api-access-2cpv8\") pod \"mariadb-operator-controller-manager-54b5986bb8-g9kpn\" (UID: \"73a88858-0760-4b40-b57f-71e4f9977129\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-g9kpn" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.683156 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-2chns" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.690917 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv77l\" (UniqueName: \"kubernetes.io/projected/38f287c7-6c43-4dc9-b547-0d494704874a-kube-api-access-fv77l\") pod \"manila-operator-controller-manager-58f887965d-mf8r8\" (UID: \"38f287c7-6c43-4dc9-b547-0d494704874a\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.700047 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-58vj4"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.711687 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-58vj4"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.711794 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-58vj4" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.716885 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-6rdk9" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.722392 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fx66w\" (UniqueName: \"kubernetes.io/projected/7ea09c1e-d85f-435e-a029-8b4f0df06839-kube-api-access-fx66w\") pod \"octavia-operator-controller-manager-54cfbf4c7d-ss948\" (UID: \"7ea09c1e-d85f-435e-a029-8b4f0df06839\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-ss948" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.749210 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.750327 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.753103 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.753385 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-69pdr" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.778456 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-vxgn2"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.779523 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-vxgn2" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.782351 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-dgrnr" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.784271 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.805766 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.807055 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.809681 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-g9kpn" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.812377 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-lknd5" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.824844 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.826090 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-vxgn2"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.837718 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-64dc7b4899-mmn8p"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.827460 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nl2b\" (UniqueName: \"kubernetes.io/projected/a9a5349a-868d-4c3e-a3ab-f57b55643759-kube-api-access-8nl2b\") pod \"nova-operator-controller-manager-cfbb9c588-58vj4\" (UID: \"a9a5349a-868d-4c3e-a3ab-f57b55643759\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-58vj4" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.838657 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fx66w\" (UniqueName: \"kubernetes.io/projected/7ea09c1e-d85f-435e-a029-8b4f0df06839-kube-api-access-fx66w\") pod \"octavia-operator-controller-manager-54cfbf4c7d-ss948\" (UID: \"7ea09c1e-d85f-435e-a029-8b4f0df06839\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-ss948" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.826592 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.831720 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-dc68g" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.844406 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-g4j88" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.845997 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.847104 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.847182 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.848301 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-64dc7b4899-mmn8p" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.856534 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.860217 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-7wjtl" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.860652 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.864077 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-65hg5" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.872899 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.880424 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-64dc7b4899-mmn8p"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.900272 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.902351 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.903597 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.905219 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-qt9dd" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.909740 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fx66w\" (UniqueName: \"kubernetes.io/projected/7ea09c1e-d85f-435e-a029-8b4f0df06839-kube-api-access-fx66w\") pod \"octavia-operator-controller-manager-54cfbf4c7d-ss948\" (UID: \"7ea09c1e-d85f-435e-a029-8b4f0df06839\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-ss948" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.913398 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq"] Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.942150 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a491c1c-85db-4efd-b944-40b1651a3c18-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d\" (UID: \"4a491c1c-85db-4efd-b944-40b1651a3c18\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.942746 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snlwt\" (UniqueName: \"kubernetes.io/projected/cfd203de-6a3e-4fdf-af55-3f5eaa62c5c7-kube-api-access-snlwt\") pod \"ovn-operator-controller-manager-54fc5f65b7-vxgn2\" (UID: \"cfd203de-6a3e-4fdf-af55-3f5eaa62c5c7\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-vxgn2" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.942806 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78vnk\" (UniqueName: \"kubernetes.io/projected/ef1391a6-60c0-47c9-aa4f-7fa39f90d55b-kube-api-access-78vnk\") pod \"placement-operator-controller-manager-5b797b8dff-hmsp7\" (UID: \"ef1391a6-60c0-47c9-aa4f-7fa39f90d55b\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.942832 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m225\" (UniqueName: \"kubernetes.io/projected/c847a619-1224-4a3c-bc57-bb75924d8553-kube-api-access-7m225\") pod \"swift-operator-controller-manager-d656998f4-q2p2n\" (UID: \"c847a619-1224-4a3c-bc57-bb75924d8553\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.942869 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nl2b\" (UniqueName: \"kubernetes.io/projected/a9a5349a-868d-4c3e-a3ab-f57b55643759-kube-api-access-8nl2b\") pod \"nova-operator-controller-manager-cfbb9c588-58vj4\" (UID: \"a9a5349a-868d-4c3e-a3ab-f57b55643759\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-58vj4" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.942929 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87zdc\" (UniqueName: \"kubernetes.io/projected/4a491c1c-85db-4efd-b944-40b1651a3c18-kube-api-access-87zdc\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d\" (UID: \"4a491c1c-85db-4efd-b944-40b1651a3c18\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.952917 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-ss948" Nov 24 17:16:30 crc kubenswrapper[4777]: I1124 17:16:30.987910 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nl2b\" (UniqueName: \"kubernetes.io/projected/a9a5349a-868d-4c3e-a3ab-f57b55643759-kube-api-access-8nl2b\") pod \"nova-operator-controller-manager-cfbb9c588-58vj4\" (UID: \"a9a5349a-868d-4c3e-a3ab-f57b55643759\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-58vj4" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.003735 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8"] Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.006994 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.019393 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.019672 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-bp4d5" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.020274 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8"] Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.043635 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtq6s\" (UniqueName: \"kubernetes.io/projected/2f5425e7-4f78-4b35-bb3f-b1c0065cfc42-kube-api-access-xtq6s\") pod \"test-operator-controller-manager-b4c496f69-jbqdt\" (UID: \"2f5425e7-4f78-4b35-bb3f-b1c0065cfc42\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.043702 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59dj8\" (UniqueName: \"kubernetes.io/projected/6f802034-e827-4870-9e32-d1d0cc263587-kube-api-access-59dj8\") pod \"telemetry-operator-controller-manager-64dc7b4899-mmn8p\" (UID: \"6f802034-e827-4870-9e32-d1d0cc263587\") " pod="openstack-operators/telemetry-operator-controller-manager-64dc7b4899-mmn8p" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.043730 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87zdc\" (UniqueName: \"kubernetes.io/projected/4a491c1c-85db-4efd-b944-40b1651a3c18-kube-api-access-87zdc\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d\" (UID: \"4a491c1c-85db-4efd-b944-40b1651a3c18\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.043763 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a491c1c-85db-4efd-b944-40b1651a3c18-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d\" (UID: \"4a491c1c-85db-4efd-b944-40b1651a3c18\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.043800 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snlwt\" (UniqueName: \"kubernetes.io/projected/cfd203de-6a3e-4fdf-af55-3f5eaa62c5c7-kube-api-access-snlwt\") pod \"ovn-operator-controller-manager-54fc5f65b7-vxgn2\" (UID: \"cfd203de-6a3e-4fdf-af55-3f5eaa62c5c7\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-vxgn2" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.043838 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78vnk\" (UniqueName: \"kubernetes.io/projected/ef1391a6-60c0-47c9-aa4f-7fa39f90d55b-kube-api-access-78vnk\") pod \"placement-operator-controller-manager-5b797b8dff-hmsp7\" (UID: \"ef1391a6-60c0-47c9-aa4f-7fa39f90d55b\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.043860 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m225\" (UniqueName: \"kubernetes.io/projected/c847a619-1224-4a3c-bc57-bb75924d8553-kube-api-access-7m225\") pod \"swift-operator-controller-manager-d656998f4-q2p2n\" (UID: \"c847a619-1224-4a3c-bc57-bb75924d8553\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.043891 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttpsl\" (UniqueName: \"kubernetes.io/projected/44e68f24-0747-476d-b3e5-11e2952c3792-kube-api-access-ttpsl\") pod \"watcher-operator-controller-manager-8c6448b9f-g5jdq\" (UID: \"44e68f24-0747-476d-b3e5-11e2952c3792\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq" Nov 24 17:16:31 crc kubenswrapper[4777]: E1124 17:16:31.044359 4777 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 17:16:31 crc kubenswrapper[4777]: E1124 17:16:31.044401 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4a491c1c-85db-4efd-b944-40b1651a3c18-cert podName:4a491c1c-85db-4efd-b944-40b1651a3c18 nodeName:}" failed. No retries permitted until 2025-11-24 17:16:31.544385044 +0000 UTC m=+949.703580093 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4a491c1c-85db-4efd-b944-40b1651a3c18-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" (UID: "4a491c1c-85db-4efd-b944-40b1651a3c18") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.051110 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm"] Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.053471 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.056460 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm"] Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.058347 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-fnwwg" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.078702 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snlwt\" (UniqueName: \"kubernetes.io/projected/cfd203de-6a3e-4fdf-af55-3f5eaa62c5c7-kube-api-access-snlwt\") pod \"ovn-operator-controller-manager-54fc5f65b7-vxgn2\" (UID: \"cfd203de-6a3e-4fdf-af55-3f5eaa62c5c7\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-vxgn2" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.083312 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-58vj4" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.087864 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78vnk\" (UniqueName: \"kubernetes.io/projected/ef1391a6-60c0-47c9-aa4f-7fa39f90d55b-kube-api-access-78vnk\") pod \"placement-operator-controller-manager-5b797b8dff-hmsp7\" (UID: \"ef1391a6-60c0-47c9-aa4f-7fa39f90d55b\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.088221 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87zdc\" (UniqueName: \"kubernetes.io/projected/4a491c1c-85db-4efd-b944-40b1651a3c18-kube-api-access-87zdc\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d\" (UID: \"4a491c1c-85db-4efd-b944-40b1651a3c18\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.088528 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m225\" (UniqueName: \"kubernetes.io/projected/c847a619-1224-4a3c-bc57-bb75924d8553-kube-api-access-7m225\") pod \"swift-operator-controller-manager-d656998f4-q2p2n\" (UID: \"c847a619-1224-4a3c-bc57-bb75924d8553\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.151149 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q46t8\" (UniqueName: \"kubernetes.io/projected/783eef36-329f-4ceb-b516-020fe4cf5151-kube-api-access-q46t8\") pod \"openstack-operator-controller-manager-7856bf9fdc-k8wq8\" (UID: \"783eef36-329f-4ceb-b516-020fe4cf5151\") " pod="openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.151296 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttpsl\" (UniqueName: \"kubernetes.io/projected/44e68f24-0747-476d-b3e5-11e2952c3792-kube-api-access-ttpsl\") pod \"watcher-operator-controller-manager-8c6448b9f-g5jdq\" (UID: \"44e68f24-0747-476d-b3e5-11e2952c3792\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.151363 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtq6s\" (UniqueName: \"kubernetes.io/projected/2f5425e7-4f78-4b35-bb3f-b1c0065cfc42-kube-api-access-xtq6s\") pod \"test-operator-controller-manager-b4c496f69-jbqdt\" (UID: \"2f5425e7-4f78-4b35-bb3f-b1c0065cfc42\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.151460 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59dj8\" (UniqueName: \"kubernetes.io/projected/6f802034-e827-4870-9e32-d1d0cc263587-kube-api-access-59dj8\") pod \"telemetry-operator-controller-manager-64dc7b4899-mmn8p\" (UID: \"6f802034-e827-4870-9e32-d1d0cc263587\") " pod="openstack-operators/telemetry-operator-controller-manager-64dc7b4899-mmn8p" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.151512 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/783eef36-329f-4ceb-b516-020fe4cf5151-cert\") pod \"openstack-operator-controller-manager-7856bf9fdc-k8wq8\" (UID: \"783eef36-329f-4ceb-b516-020fe4cf5151\") " pod="openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.192465 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtq6s\" (UniqueName: \"kubernetes.io/projected/2f5425e7-4f78-4b35-bb3f-b1c0065cfc42-kube-api-access-xtq6s\") pod \"test-operator-controller-manager-b4c496f69-jbqdt\" (UID: \"2f5425e7-4f78-4b35-bb3f-b1c0065cfc42\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.205787 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59dj8\" (UniqueName: \"kubernetes.io/projected/6f802034-e827-4870-9e32-d1d0cc263587-kube-api-access-59dj8\") pod \"telemetry-operator-controller-manager-64dc7b4899-mmn8p\" (UID: \"6f802034-e827-4870-9e32-d1d0cc263587\") " pod="openstack-operators/telemetry-operator-controller-manager-64dc7b4899-mmn8p" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.209006 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttpsl\" (UniqueName: \"kubernetes.io/projected/44e68f24-0747-476d-b3e5-11e2952c3792-kube-api-access-ttpsl\") pod \"watcher-operator-controller-manager-8c6448b9f-g5jdq\" (UID: \"44e68f24-0747-476d-b3e5-11e2952c3792\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.253180 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/783eef36-329f-4ceb-b516-020fe4cf5151-cert\") pod \"openstack-operator-controller-manager-7856bf9fdc-k8wq8\" (UID: \"783eef36-329f-4ceb-b516-020fe4cf5151\") " pod="openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.253269 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q46t8\" (UniqueName: \"kubernetes.io/projected/783eef36-329f-4ceb-b516-020fe4cf5151-kube-api-access-q46t8\") pod \"openstack-operator-controller-manager-7856bf9fdc-k8wq8\" (UID: \"783eef36-329f-4ceb-b516-020fe4cf5151\") " pod="openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.253334 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzpfj\" (UniqueName: \"kubernetes.io/projected/386e5031-5136-4c3f-9ba2-6051496e0ed9-kube-api-access-qzpfj\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm\" (UID: \"386e5031-5136-4c3f-9ba2-6051496e0ed9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm" Nov 24 17:16:31 crc kubenswrapper[4777]: E1124 17:16:31.253579 4777 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 24 17:16:31 crc kubenswrapper[4777]: E1124 17:16:31.253631 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/783eef36-329f-4ceb-b516-020fe4cf5151-cert podName:783eef36-329f-4ceb-b516-020fe4cf5151 nodeName:}" failed. No retries permitted until 2025-11-24 17:16:31.753613877 +0000 UTC m=+949.912808926 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/783eef36-329f-4ceb-b516-020fe4cf5151-cert") pod "openstack-operator-controller-manager-7856bf9fdc-k8wq8" (UID: "783eef36-329f-4ceb-b516-020fe4cf5151") : secret "webhook-server-cert" not found Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.254229 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-vxgn2" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.284305 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q46t8\" (UniqueName: \"kubernetes.io/projected/783eef36-329f-4ceb-b516-020fe4cf5151-kube-api-access-q46t8\") pod \"openstack-operator-controller-manager-7856bf9fdc-k8wq8\" (UID: \"783eef36-329f-4ceb-b516-020fe4cf5151\") " pod="openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.291706 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.320820 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.357831 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzpfj\" (UniqueName: \"kubernetes.io/projected/386e5031-5136-4c3f-9ba2-6051496e0ed9-kube-api-access-qzpfj\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm\" (UID: \"386e5031-5136-4c3f-9ba2-6051496e0ed9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.370028 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-zmms9"] Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.377762 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-82ppc"] Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.379449 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzpfj\" (UniqueName: \"kubernetes.io/projected/386e5031-5136-4c3f-9ba2-6051496e0ed9-kube-api-access-qzpfj\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm\" (UID: \"386e5031-5136-4c3f-9ba2-6051496e0ed9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.392221 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-64dc7b4899-mmn8p" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.423496 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.451267 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.501487 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.502348 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-wmvls"] Nov 24 17:16:31 crc kubenswrapper[4777]: W1124 17:16:31.526847 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7f44c61_027a_41c4_9665_8394d579ba33.slice/crio-3049bd1132ce308422f848acff40829983a9aac36c6d0bc69fa95dcab59877f5 WatchSource:0}: Error finding container 3049bd1132ce308422f848acff40829983a9aac36c6d0bc69fa95dcab59877f5: Status 404 returned error can't find the container with id 3049bd1132ce308422f848acff40829983a9aac36c6d0bc69fa95dcab59877f5 Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.553893 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-klpxr"] Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.560602 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a491c1c-85db-4efd-b944-40b1651a3c18-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d\" (UID: \"4a491c1c-85db-4efd-b944-40b1651a3c18\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.563603 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a491c1c-85db-4efd-b944-40b1651a3c18-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d\" (UID: \"4a491c1c-85db-4efd-b944-40b1651a3c18\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.674039 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-9mzvw"] Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.770767 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/783eef36-329f-4ceb-b516-020fe4cf5151-cert\") pod \"openstack-operator-controller-manager-7856bf9fdc-k8wq8\" (UID: \"783eef36-329f-4ceb-b516-020fe4cf5151\") " pod="openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.774003 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/783eef36-329f-4ceb-b516-020fe4cf5151-cert\") pod \"openstack-operator-controller-manager-7856bf9fdc-k8wq8\" (UID: \"783eef36-329f-4ceb-b516-020fe4cf5151\") " pod="openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.774303 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.774689 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-2chns"] Nov 24 17:16:31 crc kubenswrapper[4777]: W1124 17:16:31.777957 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd1db7a5_22e3_4f16_a026_0b8895ef292d.slice/crio-33725a932060455ddae52b6df0de7b37a96e655f31b3f81f5641de30469df35a WatchSource:0}: Error finding container 33725a932060455ddae52b6df0de7b37a96e655f31b3f81f5641de30469df35a: Status 404 returned error can't find the container with id 33725a932060455ddae52b6df0de7b37a96e655f31b3f81f5641de30469df35a Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.780731 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-dc68g"] Nov 24 17:16:31 crc kubenswrapper[4777]: W1124 17:16:31.791256 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d2ecc53_902f_41f3_80f4_93f9e48c8532.slice/crio-c3eec8d062c5a2b6be9291c64a20b4158c58d4b5b8c4b37e275419ca5aa8aa3f WatchSource:0}: Error finding container c3eec8d062c5a2b6be9291c64a20b4158c58d4b5b8c4b37e275419ca5aa8aa3f: Status 404 returned error can't find the container with id c3eec8d062c5a2b6be9291c64a20b4158c58d4b5b8c4b37e275419ca5aa8aa3f Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.836733 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" Nov 24 17:16:31 crc kubenswrapper[4777]: I1124 17:16:31.941810 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-g9kpn"] Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.036247 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-ss948"] Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.046746 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-vxgn2"] Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.070332 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8"] Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.074057 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc"] Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.077754 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-58vj4"] Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.083221 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-6xxk9"] Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.084750 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx"] Nov 24 17:16:32 crc kubenswrapper[4777]: W1124 17:16:32.085006 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8401afce_ddb4_4195_b6ae_7e5c91128525.slice/crio-f318f34aa31f503503baec474ddaee17e80fd81a6daf60a6fa7b41f321def82e WatchSource:0}: Error finding container f318f34aa31f503503baec474ddaee17e80fd81a6daf60a6fa7b41f321def82e: Status 404 returned error can't find the container with id f318f34aa31f503503baec474ddaee17e80fd81a6daf60a6fa7b41f321def82e Nov 24 17:16:32 crc kubenswrapper[4777]: W1124 17:16:32.088490 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9a5349a_868d_4c3e_a3ab_f57b55643759.slice/crio-dd25f3c671dcd90b6010864c1473a198240204e53febf48b4617d9a729721733 WatchSource:0}: Error finding container dd25f3c671dcd90b6010864c1473a198240204e53febf48b4617d9a729721733: Status 404 returned error can't find the container with id dd25f3c671dcd90b6010864c1473a198240204e53febf48b4617d9a729721733 Nov 24 17:16:32 crc kubenswrapper[4777]: W1124 17:16:32.097275 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38f287c7_6c43_4dc9_b547_0d494704874a.slice/crio-7fb27c5ba5367dbd25ea6e8893cb973f9ae2cb8b521f4330d75fc8f04056e9af WatchSource:0}: Error finding container 7fb27c5ba5367dbd25ea6e8893cb973f9ae2cb8b521f4330d75fc8f04056e9af: Status 404 returned error can't find the container with id 7fb27c5ba5367dbd25ea6e8893cb973f9ae2cb8b521f4330d75fc8f04056e9af Nov 24 17:16:32 crc kubenswrapper[4777]: E1124 17:16:32.103125 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vrxv7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7454b96578-v5vgx_openstack-operators(f1425713-2fd6-4108-abe3-9f46063e98f5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:16:32 crc kubenswrapper[4777]: E1124 17:16:32.105407 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fv77l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-58f887965d-mf8r8_openstack-operators(38f287c7-6c43-4dc9-b547-0d494704874a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.141658 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7"] Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.174080 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n"] Nov 24 17:16:32 crc kubenswrapper[4777]: W1124 17:16:32.181521 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef1391a6_60c0_47c9_aa4f_7fa39f90d55b.slice/crio-a06ab59141d35574d379f5909f09b1655445e09b8256347d5ae2738b7e5bb81b WatchSource:0}: Error finding container a06ab59141d35574d379f5909f09b1655445e09b8256347d5ae2738b7e5bb81b: Status 404 returned error can't find the container with id a06ab59141d35574d379f5909f09b1655445e09b8256347d5ae2738b7e5bb81b Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.183264 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-64dc7b4899-mmn8p"] Nov 24 17:16:32 crc kubenswrapper[4777]: E1124 17:16:32.185838 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-78vnk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5b797b8dff-hmsp7_openstack-operators(ef1391a6-60c0-47c9-aa4f-7fa39f90d55b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.189314 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq"] Nov 24 17:16:32 crc kubenswrapper[4777]: W1124 17:16:32.198300 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc847a619_1224_4a3c_bc57_bb75924d8553.slice/crio-3fb8708499df29243a4ce4a9d48492a212748413a6a0fda8beb0a8f6ae0c9dae WatchSource:0}: Error finding container 3fb8708499df29243a4ce4a9d48492a212748413a6a0fda8beb0a8f6ae0c9dae: Status 404 returned error can't find the container with id 3fb8708499df29243a4ce4a9d48492a212748413a6a0fda8beb0a8f6ae0c9dae Nov 24 17:16:32 crc kubenswrapper[4777]: E1124 17:16:32.200764 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7m225,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d656998f4-q2p2n_openstack-operators(c847a619-1224-4a3c-bc57-bb75924d8553): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:16:32 crc kubenswrapper[4777]: W1124 17:16:32.200864 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44e68f24_0747_476d_b3e5_11e2952c3792.slice/crio-eb29d1329566280ececfd1aab86335f1f77b825f995034da8b01fef8d8436e61 WatchSource:0}: Error finding container eb29d1329566280ececfd1aab86335f1f77b825f995034da8b01fef8d8436e61: Status 404 returned error can't find the container with id eb29d1329566280ececfd1aab86335f1f77b825f995034da8b01fef8d8436e61 Nov 24 17:16:32 crc kubenswrapper[4777]: E1124 17:16:32.251755 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ttpsl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-8c6448b9f-g5jdq_openstack-operators(44e68f24-0747-476d-b3e5-11e2952c3792): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.261290 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n" event={"ID":"c847a619-1224-4a3c-bc57-bb75924d8553","Type":"ContainerStarted","Data":"3fb8708499df29243a4ce4a9d48492a212748413a6a0fda8beb0a8f6ae0c9dae"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.267215 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-dc68g" event={"ID":"fd1db7a5-22e3-4f16-a026-0b8895ef292d","Type":"ContainerStarted","Data":"33725a932060455ddae52b6df0de7b37a96e655f31b3f81f5641de30469df35a"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.269457 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc" event={"ID":"b755cc8b-dd00-4b28-8fb4-908e0e3db9a5","Type":"ContainerStarted","Data":"a760e6bc6cb3271f1dda4d990b8176502dd1f0285aa9d9db4e8af5e40a36131b"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.276349 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-2chns" event={"ID":"1d2ecc53-902f-41f3-80f4-93f9e48c8532","Type":"ContainerStarted","Data":"c3eec8d062c5a2b6be9291c64a20b4158c58d4b5b8c4b37e275419ca5aa8aa3f"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.277832 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-82ppc" event={"ID":"a7457835-23e4-460e-a155-07c1d4e7c30e","Type":"ContainerStarted","Data":"51c5024cd4296fccd2112ad4cedb7926d09d10fda6b484fe98e6e465345bcbd5"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.279818 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-vxgn2" event={"ID":"cfd203de-6a3e-4fdf-af55-3f5eaa62c5c7","Type":"ContainerStarted","Data":"4507f882cd7f8aab8aede774b47748ac1d13c9462b48824c89679e7dc165ccbc"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.281275 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-ss948" event={"ID":"7ea09c1e-d85f-435e-a029-8b4f0df06839","Type":"ContainerStarted","Data":"3e8321fc23567be89de263df18e12e58fd392fa58ec7281dc6687022d12ba1f0"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.291726 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-g9kpn" event={"ID":"73a88858-0760-4b40-b57f-71e4f9977129","Type":"ContainerStarted","Data":"85af58db7c2bd08efcde306028cf5f0e3591539ae44188878ba7f9a3b2ef51a2"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.296923 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-6xxk9" event={"ID":"8401afce-ddb4-4195-b6ae-7e5c91128525","Type":"ContainerStarted","Data":"f318f34aa31f503503baec474ddaee17e80fd81a6daf60a6fa7b41f321def82e"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.301867 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-klpxr" event={"ID":"99bfc986-b80d-4193-bf32-4488c11f0066","Type":"ContainerStarted","Data":"6da8885995cbbcfba21753946a0a4008223e809646a32632415e3e4ea00e1f71"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.303022 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8" event={"ID":"38f287c7-6c43-4dc9-b547-0d494704874a","Type":"ContainerStarted","Data":"7fb27c5ba5367dbd25ea6e8893cb973f9ae2cb8b521f4330d75fc8f04056e9af"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.304096 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-9mzvw" event={"ID":"fe817743-7f1e-46a4-9757-5b678ddc097a","Type":"ContainerStarted","Data":"5cb50039cfd64ae26039321742d3400b6e3ca3a17332c3019a015bf43beff9ac"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.304920 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx" event={"ID":"f1425713-2fd6-4108-abe3-9f46063e98f5","Type":"ContainerStarted","Data":"4c3453419d44053929e25d6158a2c398511eacd7a717d1df606df05a887f8381"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.305913 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq" event={"ID":"44e68f24-0747-476d-b3e5-11e2952c3792","Type":"ContainerStarted","Data":"eb29d1329566280ececfd1aab86335f1f77b825f995034da8b01fef8d8436e61"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.307672 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-zmms9" event={"ID":"1d3fc8a3-f3cf-4f37-846d-b244a52415f8","Type":"ContainerStarted","Data":"a6f6563f4052560203400c65a1e048bb9ac339876d5ddbed085f15b5fb2635ee"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.309750 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-58vj4" event={"ID":"a9a5349a-868d-4c3e-a3ab-f57b55643759","Type":"ContainerStarted","Data":"dd25f3c671dcd90b6010864c1473a198240204e53febf48b4617d9a729721733"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.316222 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-wmvls" event={"ID":"f7f44c61-027a-41c4-9665-8394d579ba33","Type":"ContainerStarted","Data":"3049bd1132ce308422f848acff40829983a9aac36c6d0bc69fa95dcab59877f5"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.318439 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7" event={"ID":"ef1391a6-60c0-47c9-aa4f-7fa39f90d55b","Type":"ContainerStarted","Data":"a06ab59141d35574d379f5909f09b1655445e09b8256347d5ae2738b7e5bb81b"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.320466 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-64dc7b4899-mmn8p" event={"ID":"6f802034-e827-4870-9e32-d1d0cc263587","Type":"ContainerStarted","Data":"984f3a1e86cbb2fadcac54ea55b68e27c1a5142a31708c3c75bdcbd8be3b8d1b"} Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.375515 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt"] Nov 24 17:16:32 crc kubenswrapper[4777]: E1124 17:16:32.398641 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xtq6s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-b4c496f69-jbqdt_openstack-operators(2f5425e7-4f78-4b35-bb3f-b1c0065cfc42): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.417081 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm"] Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.436845 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8"] Nov 24 17:16:32 crc kubenswrapper[4777]: I1124 17:16:32.508944 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d"] Nov 24 17:16:32 crc kubenswrapper[4777]: W1124 17:16:32.537424 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a491c1c_85db_4efd_b944_40b1651a3c18.slice/crio-883d2a1afc09eaf639e3a22d9bbcb358ac5a8fa87f8feabd4cf2e43346112f44 WatchSource:0}: Error finding container 883d2a1afc09eaf639e3a22d9bbcb358ac5a8fa87f8feabd4cf2e43346112f44: Status 404 returned error can't find the container with id 883d2a1afc09eaf639e3a22d9bbcb358ac5a8fa87f8feabd4cf2e43346112f44 Nov 24 17:16:32 crc kubenswrapper[4777]: E1124 17:16:32.546327 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-87zdc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d_openstack-operators(4a491c1c-85db-4efd-b944-40b1651a3c18): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 17:16:32 crc kubenswrapper[4777]: E1124 17:16:32.934756 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx" podUID="f1425713-2fd6-4108-abe3-9f46063e98f5" Nov 24 17:16:32 crc kubenswrapper[4777]: E1124 17:16:32.937598 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8" podUID="38f287c7-6c43-4dc9-b547-0d494704874a" Nov 24 17:16:32 crc kubenswrapper[4777]: E1124 17:16:32.974343 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n" podUID="c847a619-1224-4a3c-bc57-bb75924d8553" Nov 24 17:16:32 crc kubenswrapper[4777]: E1124 17:16:32.974496 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7" podUID="ef1391a6-60c0-47c9-aa4f-7fa39f90d55b" Nov 24 17:16:32 crc kubenswrapper[4777]: E1124 17:16:32.974580 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq" podUID="44e68f24-0747-476d-b3e5-11e2952c3792" Nov 24 17:16:32 crc kubenswrapper[4777]: E1124 17:16:32.990972 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" podUID="4a491c1c-85db-4efd-b944-40b1651a3c18" Nov 24 17:16:33 crc kubenswrapper[4777]: E1124 17:16:33.213091 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt" podUID="2f5425e7-4f78-4b35-bb3f-b1c0065cfc42" Nov 24 17:16:33 crc kubenswrapper[4777]: I1124 17:16:33.339503 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n" event={"ID":"c847a619-1224-4a3c-bc57-bb75924d8553","Type":"ContainerStarted","Data":"7d60497296e07ddb7384541fb8ec0f1ed645f17725ae33f18c79daa5f3fd71ad"} Nov 24 17:16:33 crc kubenswrapper[4777]: E1124 17:16:33.342188 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n" podUID="c847a619-1224-4a3c-bc57-bb75924d8553" Nov 24 17:16:33 crc kubenswrapper[4777]: I1124 17:16:33.342475 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx" event={"ID":"f1425713-2fd6-4108-abe3-9f46063e98f5","Type":"ContainerStarted","Data":"0051f97dd581eaf82fa7e6f72ba4c390ad888b483291c121f3c412b9ee0c82ee"} Nov 24 17:16:33 crc kubenswrapper[4777]: I1124 17:16:33.343694 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm" event={"ID":"386e5031-5136-4c3f-9ba2-6051496e0ed9","Type":"ContainerStarted","Data":"604cc7ef787840e73d610a9cef441b93eed509a993ba80936c29ddd36531ebaa"} Nov 24 17:16:33 crc kubenswrapper[4777]: E1124 17:16:33.343712 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx" podUID="f1425713-2fd6-4108-abe3-9f46063e98f5" Nov 24 17:16:33 crc kubenswrapper[4777]: I1124 17:16:33.344949 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" event={"ID":"4a491c1c-85db-4efd-b944-40b1651a3c18","Type":"ContainerStarted","Data":"2200ffb0d79454c7fbddfd94230493db5aa1c17fca086a3bbfbc19b174a719a7"} Nov 24 17:16:33 crc kubenswrapper[4777]: I1124 17:16:33.344985 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" event={"ID":"4a491c1c-85db-4efd-b944-40b1651a3c18","Type":"ContainerStarted","Data":"883d2a1afc09eaf639e3a22d9bbcb358ac5a8fa87f8feabd4cf2e43346112f44"} Nov 24 17:16:33 crc kubenswrapper[4777]: E1124 17:16:33.345892 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" podUID="4a491c1c-85db-4efd-b944-40b1651a3c18" Nov 24 17:16:33 crc kubenswrapper[4777]: I1124 17:16:33.356101 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt" event={"ID":"2f5425e7-4f78-4b35-bb3f-b1c0065cfc42","Type":"ContainerStarted","Data":"a22d5dd3dafd3ced09bd49a665f2d81c90126ca6c63d252a818d8883a527add5"} Nov 24 17:16:33 crc kubenswrapper[4777]: I1124 17:16:33.356135 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt" event={"ID":"2f5425e7-4f78-4b35-bb3f-b1c0065cfc42","Type":"ContainerStarted","Data":"d054d895d08e878265d51c197e01f1b3c53ca84e33f63701d1a9f95b6d95330d"} Nov 24 17:16:33 crc kubenswrapper[4777]: E1124 17:16:33.361769 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt" podUID="2f5425e7-4f78-4b35-bb3f-b1c0065cfc42" Nov 24 17:16:33 crc kubenswrapper[4777]: I1124 17:16:33.369924 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq" event={"ID":"44e68f24-0747-476d-b3e5-11e2952c3792","Type":"ContainerStarted","Data":"86a55ed36125364cb7a74ab6912d428989cbbbe6e749291fa17b2cc1e6812991"} Nov 24 17:16:33 crc kubenswrapper[4777]: E1124 17:16:33.371654 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq" podUID="44e68f24-0747-476d-b3e5-11e2952c3792" Nov 24 17:16:33 crc kubenswrapper[4777]: I1124 17:16:33.382172 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8" event={"ID":"38f287c7-6c43-4dc9-b547-0d494704874a","Type":"ContainerStarted","Data":"d1b063c2f3215777b5b7ee9906bb7a3aba589cb4db093167e7261d0f744133e7"} Nov 24 17:16:33 crc kubenswrapper[4777]: E1124 17:16:33.391853 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a\\\"\"" pod="openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8" podUID="38f287c7-6c43-4dc9-b547-0d494704874a" Nov 24 17:16:33 crc kubenswrapper[4777]: I1124 17:16:33.393932 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8" event={"ID":"783eef36-329f-4ceb-b516-020fe4cf5151","Type":"ContainerStarted","Data":"ef98dad47ff5784a4dc0906d5a86539c1f62c2d1d983ccb3f3f9051db071070b"} Nov 24 17:16:33 crc kubenswrapper[4777]: I1124 17:16:33.393981 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8" event={"ID":"783eef36-329f-4ceb-b516-020fe4cf5151","Type":"ContainerStarted","Data":"04bde9f40bd53478b50259fa0a34665e0fa32339fa033e5d568c7e8e4dc90b93"} Nov 24 17:16:33 crc kubenswrapper[4777]: I1124 17:16:33.395237 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7" event={"ID":"ef1391a6-60c0-47c9-aa4f-7fa39f90d55b","Type":"ContainerStarted","Data":"9190f3e84bfa7fee83385d2e5e399a82e257769b8872fd9c81344b013fa64e0b"} Nov 24 17:16:33 crc kubenswrapper[4777]: E1124 17:16:33.396453 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7" podUID="ef1391a6-60c0-47c9-aa4f-7fa39f90d55b" Nov 24 17:16:34 crc kubenswrapper[4777]: I1124 17:16:34.405792 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8" event={"ID":"783eef36-329f-4ceb-b516-020fe4cf5151","Type":"ContainerStarted","Data":"18f25084015fb1599b7128147a701932e4b74217e725ce784da47f500dfbba85"} Nov 24 17:16:34 crc kubenswrapper[4777]: E1124 17:16:34.408641 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq" podUID="44e68f24-0747-476d-b3e5-11e2952c3792" Nov 24 17:16:34 crc kubenswrapper[4777]: E1124 17:16:34.414353 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt" podUID="2f5425e7-4f78-4b35-bb3f-b1c0065cfc42" Nov 24 17:16:34 crc kubenswrapper[4777]: E1124 17:16:34.414399 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a\\\"\"" pod="openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8" podUID="38f287c7-6c43-4dc9-b547-0d494704874a" Nov 24 17:16:34 crc kubenswrapper[4777]: E1124 17:16:34.414414 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx" podUID="f1425713-2fd6-4108-abe3-9f46063e98f5" Nov 24 17:16:34 crc kubenswrapper[4777]: E1124 17:16:34.414449 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" podUID="4a491c1c-85db-4efd-b944-40b1651a3c18" Nov 24 17:16:34 crc kubenswrapper[4777]: E1124 17:16:34.415591 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7" podUID="ef1391a6-60c0-47c9-aa4f-7fa39f90d55b" Nov 24 17:16:34 crc kubenswrapper[4777]: E1124 17:16:34.416127 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n" podUID="c847a619-1224-4a3c-bc57-bb75924d8553" Nov 24 17:16:34 crc kubenswrapper[4777]: I1124 17:16:34.566377 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8" podStartSLOduration=4.566363561 podStartE2EDuration="4.566363561s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:16:34.56427265 +0000 UTC m=+952.723467709" watchObservedRunningTime="2025-11-24 17:16:34.566363561 +0000 UTC m=+952.725558600" Nov 24 17:16:35 crc kubenswrapper[4777]: I1124 17:16:35.416563 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8" Nov 24 17:16:41 crc kubenswrapper[4777]: I1124 17:16:41.779778 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7856bf9fdc-k8wq8" Nov 24 17:16:44 crc kubenswrapper[4777]: E1124 17:16:44.906696 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:70cce55bcf89468c5d468ca2fc317bfc3dc5f2bef1c502df9faca2eb1293ede7" Nov 24 17:16:44 crc kubenswrapper[4777]: E1124 17:16:44.907263 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:70cce55bcf89468c5d468ca2fc317bfc3dc5f2bef1c502df9faca2eb1293ede7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qs9vj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-75fb479bcc-82ppc_openstack-operators(a7457835-23e4-460e-a155-07c1d4e7c30e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:16:57 crc kubenswrapper[4777]: I1124 17:16:57.616707 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-82ppc" event={"ID":"a7457835-23e4-460e-a155-07c1d4e7c30e","Type":"ContainerStarted","Data":"91f4bc17cf4a5f4faee70a35e2f46ac16c9396224aec38fb46e38b397719d7d6"} Nov 24 17:16:57 crc kubenswrapper[4777]: E1124 17:16:57.722313 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-82ppc" podUID="a7457835-23e4-460e-a155-07c1d4e7c30e" Nov 24 17:17:03 crc kubenswrapper[4777]: I1124 17:17:03.689124 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-wmvls" event={"ID":"f7f44c61-027a-41c4-9665-8394d579ba33","Type":"ContainerStarted","Data":"a87748a3e2f25a62aa75aa42add79fa460a2a173ee914c576b3d1b753e762f44"} Nov 24 17:17:03 crc kubenswrapper[4777]: I1124 17:17:03.690605 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-ss948" event={"ID":"7ea09c1e-d85f-435e-a029-8b4f0df06839","Type":"ContainerStarted","Data":"4537f0f7f8cb31248cf2c2ea3750ff78373a90c159e2da2368d8ddd63d218420"} Nov 24 17:17:03 crc kubenswrapper[4777]: I1124 17:17:03.696463 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc" event={"ID":"b755cc8b-dd00-4b28-8fb4-908e0e3db9a5","Type":"ContainerStarted","Data":"7dd65fd06b4e97c291b9132460018e0559064180bee7f3b33d540dff3141d62d"} Nov 24 17:17:03 crc kubenswrapper[4777]: I1124 17:17:03.699068 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-58vj4" event={"ID":"a9a5349a-868d-4c3e-a3ab-f57b55643759","Type":"ContainerStarted","Data":"56e8ba5c661665e257acfcf56ca7d78348f2cd022f31c03c7fdc0c824e5f2eb3"} Nov 24 17:17:03 crc kubenswrapper[4777]: I1124 17:17:03.701173 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-6xxk9" event={"ID":"8401afce-ddb4-4195-b6ae-7e5c91128525","Type":"ContainerStarted","Data":"1fdb8851afe90215db84794eb9d59b28c7022e71996dd0b45acacf871b33a284"} Nov 24 17:17:03 crc kubenswrapper[4777]: I1124 17:17:03.705002 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-vxgn2" event={"ID":"cfd203de-6a3e-4fdf-af55-3f5eaa62c5c7","Type":"ContainerStarted","Data":"06aa973e92735c0c74ccf73bc3d5d3a39c860f99d4be6a192375bd6dd32364df"} Nov 24 17:17:03 crc kubenswrapper[4777]: I1124 17:17:03.708604 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-9mzvw" event={"ID":"fe817743-7f1e-46a4-9757-5b678ddc097a","Type":"ContainerStarted","Data":"469abc865a7ad791363fc232bdb407d9333b4be236b6656e751df99ac61732a4"} Nov 24 17:17:03 crc kubenswrapper[4777]: I1124 17:17:03.714641 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-zmms9" event={"ID":"1d3fc8a3-f3cf-4f37-846d-b244a52415f8","Type":"ContainerStarted","Data":"278e3c02f38e68a8af6de099ddd79dc5a7c1d8f816b26324e0045d0fb367c4d6"} Nov 24 17:17:03 crc kubenswrapper[4777]: I1124 17:17:03.721721 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-klpxr" event={"ID":"99bfc986-b80d-4193-bf32-4488c11f0066","Type":"ContainerStarted","Data":"3c366290d2c9d6c01e386e470dd29f554cdc5eb9a0dc1e12a16920e9db5ca868"} Nov 24 17:17:03 crc kubenswrapper[4777]: I1124 17:17:03.723279 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm" event={"ID":"386e5031-5136-4c3f-9ba2-6051496e0ed9","Type":"ContainerStarted","Data":"2aa2ec888dabf18f5e2f2d76f866b0435bf7375605f1033c441b1053669a572d"} Nov 24 17:17:03 crc kubenswrapper[4777]: I1124 17:17:03.724821 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-2chns" event={"ID":"1d2ecc53-902f-41f3-80f4-93f9e48c8532","Type":"ContainerStarted","Data":"0788742ff43bbf2b5302d1e2bde010dfe7397d6509623d9e3bdd5478ddf6062d"} Nov 24 17:17:03 crc kubenswrapper[4777]: I1124 17:17:03.726934 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-dc68g" event={"ID":"fd1db7a5-22e3-4f16-a026-0b8895ef292d","Type":"ContainerStarted","Data":"0aadb0b773956afc1ad58e6bcd23d48ee41e15d728a952fd37be0e7c755c2920"} Nov 24 17:17:03 crc kubenswrapper[4777]: I1124 17:17:03.731280 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-64dc7b4899-mmn8p" event={"ID":"6f802034-e827-4870-9e32-d1d0cc263587","Type":"ContainerStarted","Data":"26b09a550e6732cc315915268bc3d28160e0f9809588ddcbec03721faa8827de"} Nov 24 17:17:03 crc kubenswrapper[4777]: I1124 17:17:03.741289 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm" podStartSLOduration=4.5719510119999995 podStartE2EDuration="33.741273077s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:32.476141538 +0000 UTC m=+950.635336587" lastFinishedPulling="2025-11-24 17:17:01.645463573 +0000 UTC m=+979.804658652" observedRunningTime="2025-11-24 17:17:03.736397576 +0000 UTC m=+981.895592625" watchObservedRunningTime="2025-11-24 17:17:03.741273077 +0000 UTC m=+981.900468126" Nov 24 17:17:04 crc kubenswrapper[4777]: I1124 17:17:04.747338 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7" event={"ID":"ef1391a6-60c0-47c9-aa4f-7fa39f90d55b","Type":"ContainerStarted","Data":"e3b45c39cb84149af0fe681e6a7957e10aa42de77ab2c64986a9ea368d1fa9b1"} Nov 24 17:17:04 crc kubenswrapper[4777]: I1124 17:17:04.748293 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7" Nov 24 17:17:04 crc kubenswrapper[4777]: I1124 17:17:04.778535 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7" podStartSLOduration=4.590956851 podStartE2EDuration="34.778512736s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:32.185698584 +0000 UTC m=+950.344893633" lastFinishedPulling="2025-11-24 17:17:02.373254439 +0000 UTC m=+980.532449518" observedRunningTime="2025-11-24 17:17:04.775696495 +0000 UTC m=+982.934891544" watchObservedRunningTime="2025-11-24 17:17:04.778512736 +0000 UTC m=+982.937707825" Nov 24 17:17:05 crc kubenswrapper[4777]: I1124 17:17:05.757289 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8" event={"ID":"38f287c7-6c43-4dc9-b547-0d494704874a","Type":"ContainerStarted","Data":"37641ff564ec91ef91108c171f5748c47418f6f08fad74b409c53e6679e4c282"} Nov 24 17:17:05 crc kubenswrapper[4777]: I1124 17:17:05.757879 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8" Nov 24 17:17:05 crc kubenswrapper[4777]: I1124 17:17:05.760671 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" event={"ID":"4a491c1c-85db-4efd-b944-40b1651a3c18","Type":"ContainerStarted","Data":"55dce1795106cf8deced58e99ae08fb7a7c03635073f5f4696e4097e40a5bc8c"} Nov 24 17:17:05 crc kubenswrapper[4777]: I1124 17:17:05.762912 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx" event={"ID":"f1425713-2fd6-4108-abe3-9f46063e98f5","Type":"ContainerStarted","Data":"3e372081025d9d5fe70d4286f4ea7fd1f489a2f7149cdfae44ae2cd5c3ec585b"} Nov 24 17:17:05 crc kubenswrapper[4777]: I1124 17:17:05.784481 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8" podStartSLOduration=5.495240094 podStartE2EDuration="35.78446283s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:32.105234927 +0000 UTC m=+950.264429976" lastFinishedPulling="2025-11-24 17:17:02.394457633 +0000 UTC m=+980.553652712" observedRunningTime="2025-11-24 17:17:05.780027132 +0000 UTC m=+983.939222221" watchObservedRunningTime="2025-11-24 17:17:05.78446283 +0000 UTC m=+983.943657879" Nov 24 17:17:06 crc kubenswrapper[4777]: I1124 17:17:06.775595 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-g9kpn" event={"ID":"73a88858-0760-4b40-b57f-71e4f9977129","Type":"ContainerStarted","Data":"cba0ee4b04f540eeefbef2fcbc79c7bff4fed2e4f5605f5b5f5340ee1dbcf8e9"} Nov 24 17:17:06 crc kubenswrapper[4777]: I1124 17:17:06.778687 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-58vj4" event={"ID":"a9a5349a-868d-4c3e-a3ab-f57b55643759","Type":"ContainerStarted","Data":"50b5e67f6be5ba2ab991033c0af6895ed15e304750270eecb907792c2193f2a7"} Nov 24 17:17:06 crc kubenswrapper[4777]: I1124 17:17:06.784559 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-wmvls" event={"ID":"f7f44c61-027a-41c4-9665-8394d579ba33","Type":"ContainerStarted","Data":"d9738078bb9dbfcd4dfdbf3c60a866f105fb6c00f621436a1aa1266d8225c532"} Nov 24 17:17:06 crc kubenswrapper[4777]: I1124 17:17:06.787790 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt" event={"ID":"2f5425e7-4f78-4b35-bb3f-b1c0065cfc42","Type":"ContainerStarted","Data":"55d4fd50d524d084e3e6a385b0010eb6a377aeb42c114b8a822e3125c3b218f7"} Nov 24 17:17:06 crc kubenswrapper[4777]: I1124 17:17:06.796975 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-64dc7b4899-mmn8p" event={"ID":"6f802034-e827-4870-9e32-d1d0cc263587","Type":"ContainerStarted","Data":"575a4fc6f5dd1cbceb600423a4ed2b84fe091a239c24eeacbb2c410575418a9b"} Nov 24 17:17:06 crc kubenswrapper[4777]: I1124 17:17:06.825094 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx" podStartSLOduration=6.575281791 podStartE2EDuration="36.825076716s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:32.102884019 +0000 UTC m=+950.262079068" lastFinishedPulling="2025-11-24 17:17:02.352678904 +0000 UTC m=+980.511873993" observedRunningTime="2025-11-24 17:17:06.822516982 +0000 UTC m=+984.981712091" watchObservedRunningTime="2025-11-24 17:17:06.825076716 +0000 UTC m=+984.984271775" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.817559 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-6xxk9" event={"ID":"8401afce-ddb4-4195-b6ae-7e5c91128525","Type":"ContainerStarted","Data":"2ebaaf081715dfdd6875bb99ea75e452320fe7f24b61aa725227d87a1e3c44be"} Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.817876 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-6xxk9" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.819994 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-2chns" event={"ID":"1d2ecc53-902f-41f3-80f4-93f9e48c8532","Type":"ContainerStarted","Data":"c39b2db3c8db0c6c74bfa5fd75739bc8d5c97a576d5ad633f4b12140298481d7"} Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.822421 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-9mzvw" event={"ID":"fe817743-7f1e-46a4-9757-5b678ddc097a","Type":"ContainerStarted","Data":"90de55f874bff26604456b7844832a0019199245ea9cd0d51d82758b8c377760"} Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.822837 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7969689c84-9mzvw" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.828865 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n" event={"ID":"c847a619-1224-4a3c-bc57-bb75924d8553","Type":"ContainerStarted","Data":"cc45c4762ce7602a5bc3528f02233e8db658eeae2ee51ea6ece7eb52353831bc"} Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.829302 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.830713 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-dc68g" event={"ID":"fd1db7a5-22e3-4f16-a026-0b8895ef292d","Type":"ContainerStarted","Data":"b948c6058998ee4d405db92237a63cbe51d3e71b46419aff2cc59f3f8aaf4ea0"} Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.831086 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-dc68g" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.833597 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-82ppc" event={"ID":"a7457835-23e4-460e-a155-07c1d4e7c30e","Type":"ContainerStarted","Data":"8e4e566d0b1cb4e011282d0bbe1d276c0191d2ad39ceb4fe97d095ec0794e8af"} Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.833951 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-82ppc" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.835713 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-6xxk9" podStartSLOduration=16.056774405 podStartE2EDuration="37.835683205s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:32.08667992 +0000 UTC m=+950.245874969" lastFinishedPulling="2025-11-24 17:16:53.86558871 +0000 UTC m=+972.024783769" observedRunningTime="2025-11-24 17:17:07.83309071 +0000 UTC m=+985.992285749" watchObservedRunningTime="2025-11-24 17:17:07.835683205 +0000 UTC m=+985.994878244" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.835818 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-ss948" event={"ID":"7ea09c1e-d85f-435e-a029-8b4f0df06839","Type":"ContainerStarted","Data":"0c8b359d7ba823b94fb68ee40cb72e2cb7842fe084b882f88c095eedeb57adf0"} Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.836175 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-ss948" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.840899 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc" event={"ID":"b755cc8b-dd00-4b28-8fb4-908e0e3db9a5","Type":"ContainerStarted","Data":"9cf61c2e26b0ca7f8fb3b41a9df0d5fcf93516bd8360c2bacb1c40458bad1066"} Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.844202 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-vxgn2" event={"ID":"cfd203de-6a3e-4fdf-af55-3f5eaa62c5c7","Type":"ContainerStarted","Data":"8eab53fd1cb0383a4014e34053ae61b46005383030ab764ef3fe4fb4aa618c18"} Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.845669 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-zmms9" event={"ID":"1d3fc8a3-f3cf-4f37-846d-b244a52415f8","Type":"ContainerStarted","Data":"2c3d2a7f23fc19a8d7f7c5ca35fc0f911456b03047b19f4a6c546f92dca7ac18"} Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.845831 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-zmms9" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.850423 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-klpxr" event={"ID":"99bfc986-b80d-4193-bf32-4488c11f0066","Type":"ContainerStarted","Data":"e7c3adb279d4493953f18a6f34884fbd489cc2fb7f62e57e74d63cde8efffc14"} Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.850791 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-klpxr" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.854548 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq" event={"ID":"44e68f24-0747-476d-b3e5-11e2952c3792","Type":"ContainerStarted","Data":"e974f9f1a6309ea305e1548cd321d8560489982bd7784fb436c064ef976e9b1f"} Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.854603 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.854556 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7969689c84-9mzvw" podStartSLOduration=17.342081681 podStartE2EDuration="37.854546411s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:31.697640864 +0000 UTC m=+949.856835913" lastFinishedPulling="2025-11-24 17:16:52.210105594 +0000 UTC m=+970.369300643" observedRunningTime="2025-11-24 17:17:07.8510515 +0000 UTC m=+986.010246549" watchObservedRunningTime="2025-11-24 17:17:07.854546411 +0000 UTC m=+986.013741460" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.855888 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-wmvls" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.881631 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n" podStartSLOduration=7.687479389 podStartE2EDuration="37.881616884s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:32.200653457 +0000 UTC m=+950.359848506" lastFinishedPulling="2025-11-24 17:17:02.394790912 +0000 UTC m=+980.553986001" observedRunningTime="2025-11-24 17:17:07.881419919 +0000 UTC m=+986.040614968" watchObservedRunningTime="2025-11-24 17:17:07.881616884 +0000 UTC m=+986.040811933" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.902333 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-dc68g" podStartSLOduration=8.050732239 podStartE2EDuration="37.902311793s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:31.791298584 +0000 UTC m=+949.950493623" lastFinishedPulling="2025-11-24 17:17:01.642878088 +0000 UTC m=+979.802073177" observedRunningTime="2025-11-24 17:17:07.897677879 +0000 UTC m=+986.056872928" watchObservedRunningTime="2025-11-24 17:17:07.902311793 +0000 UTC m=+986.061506842" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.916362 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-64dc7b4899-mmn8p" podStartSLOduration=20.983238236 podStartE2EDuration="37.916336099s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:32.251254551 +0000 UTC m=+950.410449600" lastFinishedPulling="2025-11-24 17:16:49.184352414 +0000 UTC m=+967.343547463" observedRunningTime="2025-11-24 17:17:07.914829715 +0000 UTC m=+986.074024764" watchObservedRunningTime="2025-11-24 17:17:07.916336099 +0000 UTC m=+986.075531158" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.940228 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-ss948" podStartSLOduration=16.725132711 podStartE2EDuration="37.940209779s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:32.044280573 +0000 UTC m=+950.203475622" lastFinishedPulling="2025-11-24 17:16:53.259357641 +0000 UTC m=+971.418552690" observedRunningTime="2025-11-24 17:17:07.935559975 +0000 UTC m=+986.094755034" watchObservedRunningTime="2025-11-24 17:17:07.940209779 +0000 UTC m=+986.099404828" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.959173 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq" podStartSLOduration=5.229883137 podStartE2EDuration="37.959147607s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:32.251602711 +0000 UTC m=+950.410797760" lastFinishedPulling="2025-11-24 17:17:04.980867171 +0000 UTC m=+983.140062230" observedRunningTime="2025-11-24 17:17:07.952251718 +0000 UTC m=+986.111446807" watchObservedRunningTime="2025-11-24 17:17:07.959147607 +0000 UTC m=+986.118342656" Nov 24 17:17:07 crc kubenswrapper[4777]: I1124 17:17:07.984731 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-58vj4" podStartSLOduration=8.434657446 podStartE2EDuration="37.984711637s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:32.091797668 +0000 UTC m=+950.250992717" lastFinishedPulling="2025-11-24 17:17:01.641851859 +0000 UTC m=+979.801046908" observedRunningTime="2025-11-24 17:17:07.979130225 +0000 UTC m=+986.138325274" watchObservedRunningTime="2025-11-24 17:17:07.984711637 +0000 UTC m=+986.143906686" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.011392 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt" podStartSLOduration=8.036858668 podStartE2EDuration="38.011366168s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:32.398512792 +0000 UTC m=+950.557707841" lastFinishedPulling="2025-11-24 17:17:02.373020262 +0000 UTC m=+980.532215341" observedRunningTime="2025-11-24 17:17:08.007114515 +0000 UTC m=+986.166309564" watchObservedRunningTime="2025-11-24 17:17:08.011366168 +0000 UTC m=+986.170561217" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.033841 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-82ppc" podStartSLOduration=4.428622586 podStartE2EDuration="38.033820438s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:31.428383704 +0000 UTC m=+949.587578753" lastFinishedPulling="2025-11-24 17:17:05.033581516 +0000 UTC m=+983.192776605" observedRunningTime="2025-11-24 17:17:08.026911108 +0000 UTC m=+986.186106157" watchObservedRunningTime="2025-11-24 17:17:08.033820438 +0000 UTC m=+986.193015487" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.041551 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-zmms9" podStartSLOduration=18.205314005 podStartE2EDuration="38.041534361s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:31.413691619 +0000 UTC m=+949.572886668" lastFinishedPulling="2025-11-24 17:16:51.249911975 +0000 UTC m=+969.409107024" observedRunningTime="2025-11-24 17:17:08.040407658 +0000 UTC m=+986.199602707" watchObservedRunningTime="2025-11-24 17:17:08.041534361 +0000 UTC m=+986.200729400" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.065458 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" podStartSLOduration=8.156751796 podStartE2EDuration="38.065434332s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:32.545866335 +0000 UTC m=+950.705061384" lastFinishedPulling="2025-11-24 17:17:02.454548841 +0000 UTC m=+980.613743920" observedRunningTime="2025-11-24 17:17:08.059858521 +0000 UTC m=+986.219053570" watchObservedRunningTime="2025-11-24 17:17:08.065434332 +0000 UTC m=+986.224629381" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.076470 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-klpxr" podStartSLOduration=16.411054225 podStartE2EDuration="38.076451521s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:31.594189591 +0000 UTC m=+949.753384640" lastFinishedPulling="2025-11-24 17:16:53.259586877 +0000 UTC m=+971.418781936" observedRunningTime="2025-11-24 17:17:08.074413402 +0000 UTC m=+986.233608451" watchObservedRunningTime="2025-11-24 17:17:08.076451521 +0000 UTC m=+986.235646570" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.096106 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-wmvls" podStartSLOduration=11.031820054 podStartE2EDuration="38.096077249s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:31.542956249 +0000 UTC m=+949.702151298" lastFinishedPulling="2025-11-24 17:16:58.607213434 +0000 UTC m=+976.766408493" observedRunningTime="2025-11-24 17:17:08.089574761 +0000 UTC m=+986.248769800" watchObservedRunningTime="2025-11-24 17:17:08.096077249 +0000 UTC m=+986.255272298" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.864898 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-g9kpn" event={"ID":"73a88858-0760-4b40-b57f-71e4f9977129","Type":"ContainerStarted","Data":"38421837e1bdf34bc53e4ef9f6448233cfee8cc6c9d8d1090afd5bd3b74b8fad"} Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.867408 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-2chns" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.869628 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7969689c84-9mzvw" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.869900 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-dc68g" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.871414 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-6xxk9" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.872033 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-wmvls" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.872107 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-ss948" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.872172 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-2chns" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.872558 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-zmms9" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.872878 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-klpxr" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.894482 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-2chns" podStartSLOduration=9.043468041 podStartE2EDuration="38.894459328s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:31.794891068 +0000 UTC m=+949.954086107" lastFinishedPulling="2025-11-24 17:17:01.645882305 +0000 UTC m=+979.805077394" observedRunningTime="2025-11-24 17:17:08.887084144 +0000 UTC m=+987.046279223" watchObservedRunningTime="2025-11-24 17:17:08.894459328 +0000 UTC m=+987.053654397" Nov 24 17:17:08 crc kubenswrapper[4777]: I1124 17:17:08.913180 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-vxgn2" podStartSLOduration=9.349463594 podStartE2EDuration="38.913152239s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:32.079127321 +0000 UTC m=+950.238322370" lastFinishedPulling="2025-11-24 17:17:01.642815906 +0000 UTC m=+979.802011015" observedRunningTime="2025-11-24 17:17:08.910530183 +0000 UTC m=+987.069725252" watchObservedRunningTime="2025-11-24 17:17:08.913152239 +0000 UTC m=+987.072347308" Nov 24 17:17:09 crc kubenswrapper[4777]: I1124 17:17:09.144505 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-g9kpn" podStartSLOduration=8.802338364 podStartE2EDuration="39.144483821s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:32.009890338 +0000 UTC m=+950.169085377" lastFinishedPulling="2025-11-24 17:17:02.352035745 +0000 UTC m=+980.511230834" observedRunningTime="2025-11-24 17:17:09.139095326 +0000 UTC m=+987.298290365" watchObservedRunningTime="2025-11-24 17:17:09.144483821 +0000 UTC m=+987.303678870" Nov 24 17:17:09 crc kubenswrapper[4777]: I1124 17:17:09.176058 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc" podStartSLOduration=18.020572741 podStartE2EDuration="39.176020184s" podCreationTimestamp="2025-11-24 17:16:30 +0000 UTC" firstStartedPulling="2025-11-24 17:16:32.101624002 +0000 UTC m=+950.260819051" lastFinishedPulling="2025-11-24 17:16:53.257071445 +0000 UTC m=+971.416266494" observedRunningTime="2025-11-24 17:17:09.161617887 +0000 UTC m=+987.320812936" watchObservedRunningTime="2025-11-24 17:17:09.176020184 +0000 UTC m=+987.335215233" Nov 24 17:17:09 crc kubenswrapper[4777]: I1124 17:17:09.872899 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-g9kpn" Nov 24 17:17:10 crc kubenswrapper[4777]: I1124 17:17:10.606957 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx" Nov 24 17:17:10 crc kubenswrapper[4777]: I1124 17:17:10.608852 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v5vgx" Nov 24 17:17:10 crc kubenswrapper[4777]: I1124 17:17:10.815877 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-g9kpn" Nov 24 17:17:10 crc kubenswrapper[4777]: I1124 17:17:10.861505 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc" Nov 24 17:17:10 crc kubenswrapper[4777]: I1124 17:17:10.870415 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-r5pxc" Nov 24 17:17:10 crc kubenswrapper[4777]: I1124 17:17:10.906244 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58f887965d-mf8r8" Nov 24 17:17:11 crc kubenswrapper[4777]: I1124 17:17:11.084553 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-58vj4" Nov 24 17:17:11 crc kubenswrapper[4777]: I1124 17:17:11.086452 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-58vj4" Nov 24 17:17:11 crc kubenswrapper[4777]: I1124 17:17:11.260061 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-vxgn2" Nov 24 17:17:11 crc kubenswrapper[4777]: I1124 17:17:11.260115 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-vxgn2" Nov 24 17:17:11 crc kubenswrapper[4777]: I1124 17:17:11.295232 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-hmsp7" Nov 24 17:17:11 crc kubenswrapper[4777]: I1124 17:17:11.324392 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d656998f4-q2p2n" Nov 24 17:17:11 crc kubenswrapper[4777]: I1124 17:17:11.393299 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-64dc7b4899-mmn8p" Nov 24 17:17:11 crc kubenswrapper[4777]: I1124 17:17:11.395416 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-64dc7b4899-mmn8p" Nov 24 17:17:11 crc kubenswrapper[4777]: I1124 17:17:11.424015 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt" Nov 24 17:17:11 crc kubenswrapper[4777]: I1124 17:17:11.425730 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-b4c496f69-jbqdt" Nov 24 17:17:11 crc kubenswrapper[4777]: I1124 17:17:11.452762 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq" Nov 24 17:17:11 crc kubenswrapper[4777]: I1124 17:17:11.457593 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-g5jdq" Nov 24 17:17:11 crc kubenswrapper[4777]: I1124 17:17:11.845958 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d" Nov 24 17:17:20 crc kubenswrapper[4777]: I1124 17:17:20.398864 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-82ppc" Nov 24 17:17:36 crc kubenswrapper[4777]: I1124 17:17:36.875676 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-zdtpv"] Nov 24 17:17:36 crc kubenswrapper[4777]: I1124 17:17:36.877501 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-zdtpv" Nov 24 17:17:36 crc kubenswrapper[4777]: I1124 17:17:36.880544 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 24 17:17:36 crc kubenswrapper[4777]: I1124 17:17:36.880559 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-shk9n" Nov 24 17:17:36 crc kubenswrapper[4777]: I1124 17:17:36.881638 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 24 17:17:36 crc kubenswrapper[4777]: I1124 17:17:36.881644 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 24 17:17:36 crc kubenswrapper[4777]: I1124 17:17:36.894095 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-zdtpv"] Nov 24 17:17:36 crc kubenswrapper[4777]: I1124 17:17:36.999468 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-v2wg2"] Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.000713 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.002800 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.017095 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-v2wg2"] Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.024413 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fea7030b-512f-4835-bde8-c9f9905aa91e-config\") pod \"dnsmasq-dns-675f4bcbfc-zdtpv\" (UID: \"fea7030b-512f-4835-bde8-c9f9905aa91e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-zdtpv" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.024531 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zkbz\" (UniqueName: \"kubernetes.io/projected/fea7030b-512f-4835-bde8-c9f9905aa91e-kube-api-access-6zkbz\") pod \"dnsmasq-dns-675f4bcbfc-zdtpv\" (UID: \"fea7030b-512f-4835-bde8-c9f9905aa91e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-zdtpv" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.125977 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zkbz\" (UniqueName: \"kubernetes.io/projected/fea7030b-512f-4835-bde8-c9f9905aa91e-kube-api-access-6zkbz\") pod \"dnsmasq-dns-675f4bcbfc-zdtpv\" (UID: \"fea7030b-512f-4835-bde8-c9f9905aa91e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-zdtpv" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.126029 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zkcf\" (UniqueName: \"kubernetes.io/projected/919dff3b-4077-4510-b4d0-d1b0080a8357-kube-api-access-6zkcf\") pod \"dnsmasq-dns-78dd6ddcc-v2wg2\" (UID: \"919dff3b-4077-4510-b4d0-d1b0080a8357\") " pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.126097 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fea7030b-512f-4835-bde8-c9f9905aa91e-config\") pod \"dnsmasq-dns-675f4bcbfc-zdtpv\" (UID: \"fea7030b-512f-4835-bde8-c9f9905aa91e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-zdtpv" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.126178 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/919dff3b-4077-4510-b4d0-d1b0080a8357-config\") pod \"dnsmasq-dns-78dd6ddcc-v2wg2\" (UID: \"919dff3b-4077-4510-b4d0-d1b0080a8357\") " pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.126229 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/919dff3b-4077-4510-b4d0-d1b0080a8357-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-v2wg2\" (UID: \"919dff3b-4077-4510-b4d0-d1b0080a8357\") " pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.127070 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fea7030b-512f-4835-bde8-c9f9905aa91e-config\") pod \"dnsmasq-dns-675f4bcbfc-zdtpv\" (UID: \"fea7030b-512f-4835-bde8-c9f9905aa91e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-zdtpv" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.143961 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zkbz\" (UniqueName: \"kubernetes.io/projected/fea7030b-512f-4835-bde8-c9f9905aa91e-kube-api-access-6zkbz\") pod \"dnsmasq-dns-675f4bcbfc-zdtpv\" (UID: \"fea7030b-512f-4835-bde8-c9f9905aa91e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-zdtpv" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.196527 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-zdtpv" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.227460 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/919dff3b-4077-4510-b4d0-d1b0080a8357-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-v2wg2\" (UID: \"919dff3b-4077-4510-b4d0-d1b0080a8357\") " pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.227752 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zkcf\" (UniqueName: \"kubernetes.io/projected/919dff3b-4077-4510-b4d0-d1b0080a8357-kube-api-access-6zkcf\") pod \"dnsmasq-dns-78dd6ddcc-v2wg2\" (UID: \"919dff3b-4077-4510-b4d0-d1b0080a8357\") " pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.228066 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/919dff3b-4077-4510-b4d0-d1b0080a8357-config\") pod \"dnsmasq-dns-78dd6ddcc-v2wg2\" (UID: \"919dff3b-4077-4510-b4d0-d1b0080a8357\") " pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.228350 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/919dff3b-4077-4510-b4d0-d1b0080a8357-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-v2wg2\" (UID: \"919dff3b-4077-4510-b4d0-d1b0080a8357\") " pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.228663 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/919dff3b-4077-4510-b4d0-d1b0080a8357-config\") pod \"dnsmasq-dns-78dd6ddcc-v2wg2\" (UID: \"919dff3b-4077-4510-b4d0-d1b0080a8357\") " pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.247786 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zkcf\" (UniqueName: \"kubernetes.io/projected/919dff3b-4077-4510-b4d0-d1b0080a8357-kube-api-access-6zkcf\") pod \"dnsmasq-dns-78dd6ddcc-v2wg2\" (UID: \"919dff3b-4077-4510-b4d0-d1b0080a8357\") " pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.322105 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.622917 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-v2wg2"] Nov 24 17:17:37 crc kubenswrapper[4777]: I1124 17:17:37.643942 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-zdtpv"] Nov 24 17:17:38 crc kubenswrapper[4777]: I1124 17:17:38.140919 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-zdtpv" event={"ID":"fea7030b-512f-4835-bde8-c9f9905aa91e","Type":"ContainerStarted","Data":"e41d41c321ea45325b88d2471a3dd8692a3f816c7ff414ece59bb088e1a622eb"} Nov 24 17:17:38 crc kubenswrapper[4777]: I1124 17:17:38.143053 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" event={"ID":"919dff3b-4077-4510-b4d0-d1b0080a8357","Type":"ContainerStarted","Data":"8e56864db48f250282015eeab268622ec6c69d1e46d6594b4c2667c976d82898"} Nov 24 17:17:39 crc kubenswrapper[4777]: I1124 17:17:39.776146 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-zdtpv"] Nov 24 17:17:39 crc kubenswrapper[4777]: I1124 17:17:39.804418 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dx7jp"] Nov 24 17:17:39 crc kubenswrapper[4777]: I1124 17:17:39.805550 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" Nov 24 17:17:39 crc kubenswrapper[4777]: I1124 17:17:39.816632 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dx7jp"] Nov 24 17:17:39 crc kubenswrapper[4777]: I1124 17:17:39.864368 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjdfq\" (UniqueName: \"kubernetes.io/projected/180cb346-ec2e-4122-8859-be44e2a33e1e-kube-api-access-gjdfq\") pod \"dnsmasq-dns-666b6646f7-dx7jp\" (UID: \"180cb346-ec2e-4122-8859-be44e2a33e1e\") " pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" Nov 24 17:17:39 crc kubenswrapper[4777]: I1124 17:17:39.864488 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/180cb346-ec2e-4122-8859-be44e2a33e1e-config\") pod \"dnsmasq-dns-666b6646f7-dx7jp\" (UID: \"180cb346-ec2e-4122-8859-be44e2a33e1e\") " pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" Nov 24 17:17:39 crc kubenswrapper[4777]: I1124 17:17:39.864523 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/180cb346-ec2e-4122-8859-be44e2a33e1e-dns-svc\") pod \"dnsmasq-dns-666b6646f7-dx7jp\" (UID: \"180cb346-ec2e-4122-8859-be44e2a33e1e\") " pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" Nov 24 17:17:39 crc kubenswrapper[4777]: I1124 17:17:39.965462 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/180cb346-ec2e-4122-8859-be44e2a33e1e-config\") pod \"dnsmasq-dns-666b6646f7-dx7jp\" (UID: \"180cb346-ec2e-4122-8859-be44e2a33e1e\") " pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" Nov 24 17:17:39 crc kubenswrapper[4777]: I1124 17:17:39.965505 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/180cb346-ec2e-4122-8859-be44e2a33e1e-dns-svc\") pod \"dnsmasq-dns-666b6646f7-dx7jp\" (UID: \"180cb346-ec2e-4122-8859-be44e2a33e1e\") " pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" Nov 24 17:17:39 crc kubenswrapper[4777]: I1124 17:17:39.965579 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjdfq\" (UniqueName: \"kubernetes.io/projected/180cb346-ec2e-4122-8859-be44e2a33e1e-kube-api-access-gjdfq\") pod \"dnsmasq-dns-666b6646f7-dx7jp\" (UID: \"180cb346-ec2e-4122-8859-be44e2a33e1e\") " pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" Nov 24 17:17:39 crc kubenswrapper[4777]: I1124 17:17:39.966439 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/180cb346-ec2e-4122-8859-be44e2a33e1e-config\") pod \"dnsmasq-dns-666b6646f7-dx7jp\" (UID: \"180cb346-ec2e-4122-8859-be44e2a33e1e\") " pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" Nov 24 17:17:39 crc kubenswrapper[4777]: I1124 17:17:39.966459 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/180cb346-ec2e-4122-8859-be44e2a33e1e-dns-svc\") pod \"dnsmasq-dns-666b6646f7-dx7jp\" (UID: \"180cb346-ec2e-4122-8859-be44e2a33e1e\") " pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" Nov 24 17:17:39 crc kubenswrapper[4777]: I1124 17:17:39.996815 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjdfq\" (UniqueName: \"kubernetes.io/projected/180cb346-ec2e-4122-8859-be44e2a33e1e-kube-api-access-gjdfq\") pod \"dnsmasq-dns-666b6646f7-dx7jp\" (UID: \"180cb346-ec2e-4122-8859-be44e2a33e1e\") " pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.093315 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-v2wg2"] Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.119817 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-j9kwb"] Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.121293 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.123836 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.147953 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-j9kwb"] Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.268493 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3370ce09-05a5-4d40-b89b-d75720b39677-config\") pod \"dnsmasq-dns-57d769cc4f-j9kwb\" (UID: \"3370ce09-05a5-4d40-b89b-d75720b39677\") " pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.268778 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3370ce09-05a5-4d40-b89b-d75720b39677-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-j9kwb\" (UID: \"3370ce09-05a5-4d40-b89b-d75720b39677\") " pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.268798 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ltzh\" (UniqueName: \"kubernetes.io/projected/3370ce09-05a5-4d40-b89b-d75720b39677-kube-api-access-2ltzh\") pod \"dnsmasq-dns-57d769cc4f-j9kwb\" (UID: \"3370ce09-05a5-4d40-b89b-d75720b39677\") " pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.370273 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3370ce09-05a5-4d40-b89b-d75720b39677-config\") pod \"dnsmasq-dns-57d769cc4f-j9kwb\" (UID: \"3370ce09-05a5-4d40-b89b-d75720b39677\") " pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.371361 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3370ce09-05a5-4d40-b89b-d75720b39677-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-j9kwb\" (UID: \"3370ce09-05a5-4d40-b89b-d75720b39677\") " pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.371391 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ltzh\" (UniqueName: \"kubernetes.io/projected/3370ce09-05a5-4d40-b89b-d75720b39677-kube-api-access-2ltzh\") pod \"dnsmasq-dns-57d769cc4f-j9kwb\" (UID: \"3370ce09-05a5-4d40-b89b-d75720b39677\") " pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.429622 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ltzh\" (UniqueName: \"kubernetes.io/projected/3370ce09-05a5-4d40-b89b-d75720b39677-kube-api-access-2ltzh\") pod \"dnsmasq-dns-57d769cc4f-j9kwb\" (UID: \"3370ce09-05a5-4d40-b89b-d75720b39677\") " pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.626701 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dx7jp"] Nov 24 17:17:40 crc kubenswrapper[4777]: W1124 17:17:40.630223 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod180cb346_ec2e_4122_8859_be44e2a33e1e.slice/crio-d234923475ff0bcc644785693bbe8e63cf649038c0de6dcbf29b34bd614da678 WatchSource:0}: Error finding container d234923475ff0bcc644785693bbe8e63cf649038c0de6dcbf29b34bd614da678: Status 404 returned error can't find the container with id d234923475ff0bcc644785693bbe8e63cf649038c0de6dcbf29b34bd614da678 Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.712678 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3370ce09-05a5-4d40-b89b-d75720b39677-config\") pod \"dnsmasq-dns-57d769cc4f-j9kwb\" (UID: \"3370ce09-05a5-4d40-b89b-d75720b39677\") " pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.713330 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3370ce09-05a5-4d40-b89b-d75720b39677-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-j9kwb\" (UID: \"3370ce09-05a5-4d40-b89b-d75720b39677\") " pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.739721 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.955299 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.956448 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.958083 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.958367 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-cks9w" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.958775 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.958899 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.960229 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.960281 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.961624 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 17:17:40 crc kubenswrapper[4777]: I1124 17:17:40.972527 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.082672 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.082747 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.082878 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.082906 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4c707277-46e4-42ea-9076-566b2d01f544-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.082930 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.082952 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-config-data\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.083006 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.083067 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ltf9\" (UniqueName: \"kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-kube-api-access-2ltf9\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.083098 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.083116 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4c707277-46e4-42ea-9076-566b2d01f544-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.083147 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.173356 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" event={"ID":"180cb346-ec2e-4122-8859-be44e2a33e1e","Type":"ContainerStarted","Data":"d234923475ff0bcc644785693bbe8e63cf649038c0de6dcbf29b34bd614da678"} Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.184345 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.184388 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.184426 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.184443 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4c707277-46e4-42ea-9076-566b2d01f544-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.184460 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.184473 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-config-data\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.184499 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.184520 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ltf9\" (UniqueName: \"kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-kube-api-access-2ltf9\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.184540 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.184556 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4c707277-46e4-42ea-9076-566b2d01f544-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.184580 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.185777 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.186848 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.187094 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.187903 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-config-data\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.188037 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.188156 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.188188 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/38e94766a61c3e0d686381901a2b5cf624719d54de394fd2452d21d58a78d20d/globalmount\"" pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.190277 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4c707277-46e4-42ea-9076-566b2d01f544-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.193454 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.193500 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4c707277-46e4-42ea-9076-566b2d01f544-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.203640 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.208213 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ltf9\" (UniqueName: \"kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-kube-api-access-2ltf9\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.228913 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-j9kwb"] Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.237825 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\") pod \"rabbitmq-server-0\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: W1124 17:17:41.238288 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3370ce09_05a5_4d40_b89b_d75720b39677.slice/crio-8005b69fc25998703c4c758ebf5d03c030e9a6be084e1edf3e5d2cb6a9713ba9 WatchSource:0}: Error finding container 8005b69fc25998703c4c758ebf5d03c030e9a6be084e1edf3e5d2cb6a9713ba9: Status 404 returned error can't find the container with id 8005b69fc25998703c4c758ebf5d03c030e9a6be084e1edf3e5d2cb6a9713ba9 Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.277062 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.284161 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.289215 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.289229 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.289274 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4hmjq" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.289292 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.289412 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.290521 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.291372 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.291599 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.332206 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.390930 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.391023 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.391071 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.391093 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbf6t\" (UniqueName: \"kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-kube-api-access-cbf6t\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.391125 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7163366e-9a39-4379-af0a-d6c7c998ab7f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.391142 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.391183 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.391199 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7163366e-9a39-4379-af0a-d6c7c998ab7f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.391222 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.391245 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.391269 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.493780 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.493901 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.493946 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.494023 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.494059 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbf6t\" (UniqueName: \"kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-kube-api-access-cbf6t\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.494085 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7163366e-9a39-4379-af0a-d6c7c998ab7f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.494112 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.494148 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.494174 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7163366e-9a39-4379-af0a-d6c7c998ab7f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.494210 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.494234 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.494814 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.495125 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.499052 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.499082 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.500140 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.500313 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.501711 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7163366e-9a39-4379-af0a-d6c7c998ab7f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.502227 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.502254 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/64e6484d692ffc427f079db163a6addcadc33dfbfa9e1c0ba7f6f979b99f192e/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.502636 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7163366e-9a39-4379-af0a-d6c7c998ab7f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.509002 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.524609 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbf6t\" (UniqueName: \"kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-kube-api-access-cbf6t\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.541518 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\") pod \"rabbitmq-cell1-server-0\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.604444 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:17:41 crc kubenswrapper[4777]: I1124 17:17:41.894756 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:17:42 crc kubenswrapper[4777]: W1124 17:17:42.052385 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7163366e_9a39_4379_af0a_d6c7c998ab7f.slice/crio-eafb2532315b1b71211c72b4081052bcf8ef5a55b922201b10bcf3e1492ba376 WatchSource:0}: Error finding container eafb2532315b1b71211c72b4081052bcf8ef5a55b922201b10bcf3e1492ba376: Status 404 returned error can't find the container with id eafb2532315b1b71211c72b4081052bcf8ef5a55b922201b10bcf3e1492ba376 Nov 24 17:17:42 crc kubenswrapper[4777]: I1124 17:17:42.055286 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:17:42 crc kubenswrapper[4777]: I1124 17:17:42.180399 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7163366e-9a39-4379-af0a-d6c7c998ab7f","Type":"ContainerStarted","Data":"eafb2532315b1b71211c72b4081052bcf8ef5a55b922201b10bcf3e1492ba376"} Nov 24 17:17:42 crc kubenswrapper[4777]: I1124 17:17:42.181607 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4c707277-46e4-42ea-9076-566b2d01f544","Type":"ContainerStarted","Data":"d8494ef2a24b9eec544579065892957d1c4431eea4c43c065715a0430702f202"} Nov 24 17:17:42 crc kubenswrapper[4777]: I1124 17:17:42.182669 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" event={"ID":"3370ce09-05a5-4d40-b89b-d75720b39677","Type":"ContainerStarted","Data":"8005b69fc25998703c4c758ebf5d03c030e9a6be084e1edf3e5d2cb6a9713ba9"} Nov 24 17:17:42 crc kubenswrapper[4777]: I1124 17:17:42.836216 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 24 17:17:42 crc kubenswrapper[4777]: I1124 17:17:42.838632 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 17:17:42 crc kubenswrapper[4777]: I1124 17:17:42.841528 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 24 17:17:42 crc kubenswrapper[4777]: I1124 17:17:42.841753 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 24 17:17:42 crc kubenswrapper[4777]: I1124 17:17:42.841856 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 24 17:17:42 crc kubenswrapper[4777]: I1124 17:17:42.844474 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-xplgq" Nov 24 17:17:42 crc kubenswrapper[4777]: I1124 17:17:42.854694 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 24 17:17:42 crc kubenswrapper[4777]: I1124 17:17:42.882418 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.018077 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w78cs\" (UniqueName: \"kubernetes.io/projected/1d71a086-37e0-4409-9db9-66e6f1374406-kube-api-access-w78cs\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.018205 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d71a086-37e0-4409-9db9-66e6f1374406-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.018252 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2380aa26-a7aa-4cf7-b1e1-0576f862988e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2380aa26-a7aa-4cf7-b1e1-0576f862988e\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.018393 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1d71a086-37e0-4409-9db9-66e6f1374406-kolla-config\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.018430 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d71a086-37e0-4409-9db9-66e6f1374406-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.018608 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1d71a086-37e0-4409-9db9-66e6f1374406-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.018657 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d71a086-37e0-4409-9db9-66e6f1374406-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.018686 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1d71a086-37e0-4409-9db9-66e6f1374406-config-data-default\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.120771 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d71a086-37e0-4409-9db9-66e6f1374406-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.120821 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1d71a086-37e0-4409-9db9-66e6f1374406-config-data-default\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.120850 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w78cs\" (UniqueName: \"kubernetes.io/projected/1d71a086-37e0-4409-9db9-66e6f1374406-kube-api-access-w78cs\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.120912 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d71a086-37e0-4409-9db9-66e6f1374406-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.120942 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2380aa26-a7aa-4cf7-b1e1-0576f862988e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2380aa26-a7aa-4cf7-b1e1-0576f862988e\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.121021 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1d71a086-37e0-4409-9db9-66e6f1374406-kolla-config\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.121049 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d71a086-37e0-4409-9db9-66e6f1374406-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.121105 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1d71a086-37e0-4409-9db9-66e6f1374406-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.121521 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1d71a086-37e0-4409-9db9-66e6f1374406-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.121891 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1d71a086-37e0-4409-9db9-66e6f1374406-config-data-default\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.122567 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1d71a086-37e0-4409-9db9-66e6f1374406-kolla-config\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.127759 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d71a086-37e0-4409-9db9-66e6f1374406-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.127832 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d71a086-37e0-4409-9db9-66e6f1374406-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.140811 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d71a086-37e0-4409-9db9-66e6f1374406-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.142066 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.142097 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2380aa26-a7aa-4cf7-b1e1-0576f862988e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2380aa26-a7aa-4cf7-b1e1-0576f862988e\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/64f3254ce03b3aa1781ab02f5066e7567d6ec738e14897dd2dc54d94ae7e5532/globalmount\"" pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.145167 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w78cs\" (UniqueName: \"kubernetes.io/projected/1d71a086-37e0-4409-9db9-66e6f1374406-kube-api-access-w78cs\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.241650 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2380aa26-a7aa-4cf7-b1e1-0576f862988e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2380aa26-a7aa-4cf7-b1e1-0576f862988e\") pod \"openstack-galera-0\" (UID: \"1d71a086-37e0-4409-9db9-66e6f1374406\") " pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.466768 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-xplgq" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.476084 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 17:17:43 crc kubenswrapper[4777]: I1124 17:17:43.954659 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 17:17:43 crc kubenswrapper[4777]: W1124 17:17:43.965036 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d71a086_37e0_4409_9db9_66e6f1374406.slice/crio-ed8fd0a64d31a2b3cf7f9895c3f090eeed71238ea0e14856f6d302a822bd0fe4 WatchSource:0}: Error finding container ed8fd0a64d31a2b3cf7f9895c3f090eeed71238ea0e14856f6d302a822bd0fe4: Status 404 returned error can't find the container with id ed8fd0a64d31a2b3cf7f9895c3f090eeed71238ea0e14856f6d302a822bd0fe4 Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.177215 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.178497 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.181274 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.181321 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.181835 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-5bzjr" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.181939 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.197198 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.203777 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1d71a086-37e0-4409-9db9-66e6f1374406","Type":"ContainerStarted","Data":"ed8fd0a64d31a2b3cf7f9895c3f090eeed71238ea0e14856f6d302a822bd0fe4"} Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.343930 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4eba3f22-b7ad-464d-a0c9-437c449d124b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.344094 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0e7f587c-a729-4bf2-8edf-6c8ec86b1a91\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e7f587c-a729-4bf2-8edf-6c8ec86b1a91\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.344131 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4eba3f22-b7ad-464d-a0c9-437c449d124b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.344256 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4eba3f22-b7ad-464d-a0c9-437c449d124b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.344310 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4eba3f22-b7ad-464d-a0c9-437c449d124b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.344382 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eba3f22-b7ad-464d-a0c9-437c449d124b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.344434 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4eba3f22-b7ad-464d-a0c9-437c449d124b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.344465 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxwxq\" (UniqueName: \"kubernetes.io/projected/4eba3f22-b7ad-464d-a0c9-437c449d124b-kube-api-access-mxwxq\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.445739 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4eba3f22-b7ad-464d-a0c9-437c449d124b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.445805 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4eba3f22-b7ad-464d-a0c9-437c449d124b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.445831 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eba3f22-b7ad-464d-a0c9-437c449d124b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.445856 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4eba3f22-b7ad-464d-a0c9-437c449d124b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.445877 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxwxq\" (UniqueName: \"kubernetes.io/projected/4eba3f22-b7ad-464d-a0c9-437c449d124b-kube-api-access-mxwxq\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.445918 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4eba3f22-b7ad-464d-a0c9-437c449d124b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.445943 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0e7f587c-a729-4bf2-8edf-6c8ec86b1a91\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e7f587c-a729-4bf2-8edf-6c8ec86b1a91\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.445991 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4eba3f22-b7ad-464d-a0c9-437c449d124b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.446778 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4eba3f22-b7ad-464d-a0c9-437c449d124b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.447063 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4eba3f22-b7ad-464d-a0c9-437c449d124b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.447502 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4eba3f22-b7ad-464d-a0c9-437c449d124b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.447999 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4eba3f22-b7ad-464d-a0c9-437c449d124b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.450828 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4eba3f22-b7ad-464d-a0c9-437c449d124b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.450846 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eba3f22-b7ad-464d-a0c9-437c449d124b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.460738 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxwxq\" (UniqueName: \"kubernetes.io/projected/4eba3f22-b7ad-464d-a0c9-437c449d124b-kube-api-access-mxwxq\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.571341 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.572301 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.576476 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.576491 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.576509 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-gj2dl" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.583870 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.649305 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0254e0e3-f9b9-4017-954a-14d3d21e37c5-kolla-config\") pod \"memcached-0\" (UID: \"0254e0e3-f9b9-4017-954a-14d3d21e37c5\") " pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.649405 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0254e0e3-f9b9-4017-954a-14d3d21e37c5-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0254e0e3-f9b9-4017-954a-14d3d21e37c5\") " pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.649574 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0254e0e3-f9b9-4017-954a-14d3d21e37c5-config-data\") pod \"memcached-0\" (UID: \"0254e0e3-f9b9-4017-954a-14d3d21e37c5\") " pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.649640 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2xb9\" (UniqueName: \"kubernetes.io/projected/0254e0e3-f9b9-4017-954a-14d3d21e37c5-kube-api-access-j2xb9\") pod \"memcached-0\" (UID: \"0254e0e3-f9b9-4017-954a-14d3d21e37c5\") " pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.649760 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0254e0e3-f9b9-4017-954a-14d3d21e37c5-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0254e0e3-f9b9-4017-954a-14d3d21e37c5\") " pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.751772 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0254e0e3-f9b9-4017-954a-14d3d21e37c5-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0254e0e3-f9b9-4017-954a-14d3d21e37c5\") " pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.751831 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0254e0e3-f9b9-4017-954a-14d3d21e37c5-config-data\") pod \"memcached-0\" (UID: \"0254e0e3-f9b9-4017-954a-14d3d21e37c5\") " pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.751854 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2xb9\" (UniqueName: \"kubernetes.io/projected/0254e0e3-f9b9-4017-954a-14d3d21e37c5-kube-api-access-j2xb9\") pod \"memcached-0\" (UID: \"0254e0e3-f9b9-4017-954a-14d3d21e37c5\") " pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.751874 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0254e0e3-f9b9-4017-954a-14d3d21e37c5-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0254e0e3-f9b9-4017-954a-14d3d21e37c5\") " pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.751937 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0254e0e3-f9b9-4017-954a-14d3d21e37c5-kolla-config\") pod \"memcached-0\" (UID: \"0254e0e3-f9b9-4017-954a-14d3d21e37c5\") " pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.752826 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0254e0e3-f9b9-4017-954a-14d3d21e37c5-kolla-config\") pod \"memcached-0\" (UID: \"0254e0e3-f9b9-4017-954a-14d3d21e37c5\") " pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.752920 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0254e0e3-f9b9-4017-954a-14d3d21e37c5-config-data\") pod \"memcached-0\" (UID: \"0254e0e3-f9b9-4017-954a-14d3d21e37c5\") " pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.755190 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0254e0e3-f9b9-4017-954a-14d3d21e37c5-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0254e0e3-f9b9-4017-954a-14d3d21e37c5\") " pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.755801 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0254e0e3-f9b9-4017-954a-14d3d21e37c5-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0254e0e3-f9b9-4017-954a-14d3d21e37c5\") " pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.775503 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2xb9\" (UniqueName: \"kubernetes.io/projected/0254e0e3-f9b9-4017-954a-14d3d21e37c5-kube-api-access-j2xb9\") pod \"memcached-0\" (UID: \"0254e0e3-f9b9-4017-954a-14d3d21e37c5\") " pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.901437 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.921953 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.922014 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0e7f587c-a729-4bf2-8edf-6c8ec86b1a91\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e7f587c-a729-4bf2-8edf-6c8ec86b1a91\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/36b2295f9daef64d62b6147531d7f060ed3e3eb2b12ed783539dc15e4d24bc72/globalmount\"" pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:44 crc kubenswrapper[4777]: I1124 17:17:44.969564 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0e7f587c-a729-4bf2-8edf-6c8ec86b1a91\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e7f587c-a729-4bf2-8edf-6c8ec86b1a91\") pod \"openstack-cell1-galera-0\" (UID: \"4eba3f22-b7ad-464d-a0c9-437c449d124b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:45 crc kubenswrapper[4777]: I1124 17:17:45.117694 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 17:17:45 crc kubenswrapper[4777]: I1124 17:17:45.435152 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 17:17:45 crc kubenswrapper[4777]: W1124 17:17:45.442255 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0254e0e3_f9b9_4017_954a_14d3d21e37c5.slice/crio-9eaed195bdea2d20481d37baae7fffd7bcdd13ca387bd6386e3766f5eed8adbc WatchSource:0}: Error finding container 9eaed195bdea2d20481d37baae7fffd7bcdd13ca387bd6386e3766f5eed8adbc: Status 404 returned error can't find the container with id 9eaed195bdea2d20481d37baae7fffd7bcdd13ca387bd6386e3766f5eed8adbc Nov 24 17:17:45 crc kubenswrapper[4777]: I1124 17:17:45.550000 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 17:17:45 crc kubenswrapper[4777]: W1124 17:17:45.568485 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4eba3f22_b7ad_464d_a0c9_437c449d124b.slice/crio-4109cc8863cd55233bcffe5bc57cd8d5d3a102c90c37caa14134a6f102827876 WatchSource:0}: Error finding container 4109cc8863cd55233bcffe5bc57cd8d5d3a102c90c37caa14134a6f102827876: Status 404 returned error can't find the container with id 4109cc8863cd55233bcffe5bc57cd8d5d3a102c90c37caa14134a6f102827876 Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.113116 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.114234 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.117743 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-p7h7g" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.126472 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.185046 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9k48\" (UniqueName: \"kubernetes.io/projected/155b4504-c156-4af8-a9e0-1dc66c9d7d43-kube-api-access-x9k48\") pod \"kube-state-metrics-0\" (UID: \"155b4504-c156-4af8-a9e0-1dc66c9d7d43\") " pod="openstack/kube-state-metrics-0" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.242345 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4eba3f22-b7ad-464d-a0c9-437c449d124b","Type":"ContainerStarted","Data":"4109cc8863cd55233bcffe5bc57cd8d5d3a102c90c37caa14134a6f102827876"} Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.243650 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0254e0e3-f9b9-4017-954a-14d3d21e37c5","Type":"ContainerStarted","Data":"9eaed195bdea2d20481d37baae7fffd7bcdd13ca387bd6386e3766f5eed8adbc"} Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.287478 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9k48\" (UniqueName: \"kubernetes.io/projected/155b4504-c156-4af8-a9e0-1dc66c9d7d43-kube-api-access-x9k48\") pod \"kube-state-metrics-0\" (UID: \"155b4504-c156-4af8-a9e0-1dc66c9d7d43\") " pod="openstack/kube-state-metrics-0" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.332730 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9k48\" (UniqueName: \"kubernetes.io/projected/155b4504-c156-4af8-a9e0-1dc66c9d7d43-kube-api-access-x9k48\") pod \"kube-state-metrics-0\" (UID: \"155b4504-c156-4af8-a9e0-1dc66c9d7d43\") " pod="openstack/kube-state-metrics-0" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.434424 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.816949 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.819063 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.825905 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.826196 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.826236 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.826370 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-qpkxk" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.826677 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.853668 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.899252 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk7m8\" (UniqueName: \"kubernetes.io/projected/055503f1-ab1c-47db-be36-37a3f572be32-kube-api-access-bk7m8\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.899451 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/055503f1-ab1c-47db-be36-37a3f572be32-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.899479 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/055503f1-ab1c-47db-be36-37a3f572be32-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.899517 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/055503f1-ab1c-47db-be36-37a3f572be32-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.899550 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/055503f1-ab1c-47db-be36-37a3f572be32-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.899684 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/055503f1-ab1c-47db-be36-37a3f572be32-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:46 crc kubenswrapper[4777]: I1124 17:17:46.899740 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/055503f1-ab1c-47db-be36-37a3f572be32-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.003415 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk7m8\" (UniqueName: \"kubernetes.io/projected/055503f1-ab1c-47db-be36-37a3f572be32-kube-api-access-bk7m8\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.003515 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/055503f1-ab1c-47db-be36-37a3f572be32-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.005201 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/055503f1-ab1c-47db-be36-37a3f572be32-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.005269 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/055503f1-ab1c-47db-be36-37a3f572be32-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.005297 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/055503f1-ab1c-47db-be36-37a3f572be32-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.005351 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/055503f1-ab1c-47db-be36-37a3f572be32-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.005366 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/055503f1-ab1c-47db-be36-37a3f572be32-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.015490 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/055503f1-ab1c-47db-be36-37a3f572be32-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.015653 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/055503f1-ab1c-47db-be36-37a3f572be32-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.016875 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/055503f1-ab1c-47db-be36-37a3f572be32-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.021096 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk7m8\" (UniqueName: \"kubernetes.io/projected/055503f1-ab1c-47db-be36-37a3f572be32-kube-api-access-bk7m8\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.021389 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/055503f1-ab1c-47db-be36-37a3f572be32-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.021531 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/055503f1-ab1c-47db-be36-37a3f572be32-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.022452 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/055503f1-ab1c-47db-be36-37a3f572be32-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"055503f1-ab1c-47db-be36-37a3f572be32\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.152826 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.410987 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.413888 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.415568 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.415889 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.415997 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-mxt75" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.416112 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.416148 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.421048 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.433874 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.516051 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-config\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.516674 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.516700 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.516757 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjm9x\" (UniqueName: \"kubernetes.io/projected/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-kube-api-access-gjm9x\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.516776 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.516794 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-872360a4-14ac-4955-839d-37d8c1b22667\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-872360a4-14ac-4955-839d-37d8c1b22667\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.516811 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.516828 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.618115 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjm9x\" (UniqueName: \"kubernetes.io/projected/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-kube-api-access-gjm9x\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.618158 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.618185 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.618210 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-872360a4-14ac-4955-839d-37d8c1b22667\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-872360a4-14ac-4955-839d-37d8c1b22667\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.618235 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.618292 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-config\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.618345 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.618369 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.620180 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.622509 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.622544 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-872360a4-14ac-4955-839d-37d8c1b22667\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-872360a4-14ac-4955-839d-37d8c1b22667\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b9e58ef1499554db3d4c243d9997bf278b9eaa38e6e18118db34549346e6e461/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.624724 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.626134 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.632830 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.635466 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.635476 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-config\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.641824 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjm9x\" (UniqueName: \"kubernetes.io/projected/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-kube-api-access-gjm9x\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.687214 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-872360a4-14ac-4955-839d-37d8c1b22667\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-872360a4-14ac-4955-839d-37d8c1b22667\") pod \"prometheus-metric-storage-0\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:47 crc kubenswrapper[4777]: I1124 17:17:47.749755 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.321771 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-q4l2k"] Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.325165 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.326992 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-dz7wm" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.328612 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.333346 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.347633 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-q4l2k"] Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.371212 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-dkbds"] Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.374736 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.383378 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-dkbds"] Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.486423 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73928e73-667c-4b69-aed9-72f32012fbdc-combined-ca-bundle\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.486496 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/73928e73-667c-4b69-aed9-72f32012fbdc-var-log-ovn\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.486548 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73928e73-667c-4b69-aed9-72f32012fbdc-scripts\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.486606 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3641f728-c179-4f10-93c2-98f74763e77f-var-lib\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.486640 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/73928e73-667c-4b69-aed9-72f32012fbdc-var-run-ovn\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.486677 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3641f728-c179-4f10-93c2-98f74763e77f-var-log\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.486715 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s7dq\" (UniqueName: \"kubernetes.io/projected/3641f728-c179-4f10-93c2-98f74763e77f-kube-api-access-8s7dq\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.486743 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/73928e73-667c-4b69-aed9-72f32012fbdc-var-run\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.486763 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3641f728-c179-4f10-93c2-98f74763e77f-etc-ovs\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.486789 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3641f728-c179-4f10-93c2-98f74763e77f-scripts\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.486811 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdwpf\" (UniqueName: \"kubernetes.io/projected/73928e73-667c-4b69-aed9-72f32012fbdc-kube-api-access-zdwpf\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.486856 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3641f728-c179-4f10-93c2-98f74763e77f-var-run\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.486899 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/73928e73-667c-4b69-aed9-72f32012fbdc-ovn-controller-tls-certs\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.588883 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73928e73-667c-4b69-aed9-72f32012fbdc-combined-ca-bundle\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.588954 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/73928e73-667c-4b69-aed9-72f32012fbdc-var-log-ovn\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.589013 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73928e73-667c-4b69-aed9-72f32012fbdc-scripts\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.589065 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3641f728-c179-4f10-93c2-98f74763e77f-var-lib\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.589094 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/73928e73-667c-4b69-aed9-72f32012fbdc-var-run-ovn\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.589119 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3641f728-c179-4f10-93c2-98f74763e77f-var-log\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.589155 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s7dq\" (UniqueName: \"kubernetes.io/projected/3641f728-c179-4f10-93c2-98f74763e77f-kube-api-access-8s7dq\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.589180 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/73928e73-667c-4b69-aed9-72f32012fbdc-var-run\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.589201 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3641f728-c179-4f10-93c2-98f74763e77f-etc-ovs\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.589220 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3641f728-c179-4f10-93c2-98f74763e77f-scripts\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.589260 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdwpf\" (UniqueName: \"kubernetes.io/projected/73928e73-667c-4b69-aed9-72f32012fbdc-kube-api-access-zdwpf\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.589296 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3641f728-c179-4f10-93c2-98f74763e77f-var-run\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.589330 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/73928e73-667c-4b69-aed9-72f32012fbdc-ovn-controller-tls-certs\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.589817 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/73928e73-667c-4b69-aed9-72f32012fbdc-var-log-ovn\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.590004 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3641f728-c179-4f10-93c2-98f74763e77f-var-log\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.590054 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3641f728-c179-4f10-93c2-98f74763e77f-etc-ovs\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.590194 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/73928e73-667c-4b69-aed9-72f32012fbdc-var-run\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.590221 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3641f728-c179-4f10-93c2-98f74763e77f-var-run\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.590344 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/73928e73-667c-4b69-aed9-72f32012fbdc-var-run-ovn\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.589987 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3641f728-c179-4f10-93c2-98f74763e77f-var-lib\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.591494 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73928e73-667c-4b69-aed9-72f32012fbdc-scripts\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.592639 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3641f728-c179-4f10-93c2-98f74763e77f-scripts\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.594235 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73928e73-667c-4b69-aed9-72f32012fbdc-combined-ca-bundle\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.596787 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/73928e73-667c-4b69-aed9-72f32012fbdc-ovn-controller-tls-certs\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.612378 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdwpf\" (UniqueName: \"kubernetes.io/projected/73928e73-667c-4b69-aed9-72f32012fbdc-kube-api-access-zdwpf\") pod \"ovn-controller-q4l2k\" (UID: \"73928e73-667c-4b69-aed9-72f32012fbdc\") " pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.617577 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s7dq\" (UniqueName: \"kubernetes.io/projected/3641f728-c179-4f10-93c2-98f74763e77f-kube-api-access-8s7dq\") pod \"ovn-controller-ovs-dkbds\" (UID: \"3641f728-c179-4f10-93c2-98f74763e77f\") " pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.667214 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q4l2k" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.742118 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.887588 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.889004 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.892507 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-dtl4r" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.893152 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.893994 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.894228 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.894382 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.914675 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.994492 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/737d5505-10bc-4e56-b846-01a3af071b38-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.994635 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/737d5505-10bc-4e56-b846-01a3af071b38-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.994661 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/737d5505-10bc-4e56-b846-01a3af071b38-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.994685 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/737d5505-10bc-4e56-b846-01a3af071b38-config\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.994736 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/737d5505-10bc-4e56-b846-01a3af071b38-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.994756 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7hk8\" (UniqueName: \"kubernetes.io/projected/737d5505-10bc-4e56-b846-01a3af071b38-kube-api-access-j7hk8\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.994797 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-63013308-26d7-4a2a-8b36-252e8653e671\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63013308-26d7-4a2a-8b36-252e8653e671\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:50 crc kubenswrapper[4777]: I1124 17:17:50.994835 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/737d5505-10bc-4e56-b846-01a3af071b38-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.096690 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/737d5505-10bc-4e56-b846-01a3af071b38-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.096824 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/737d5505-10bc-4e56-b846-01a3af071b38-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.096860 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/737d5505-10bc-4e56-b846-01a3af071b38-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.096891 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/737d5505-10bc-4e56-b846-01a3af071b38-config\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.096934 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/737d5505-10bc-4e56-b846-01a3af071b38-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.097061 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7hk8\" (UniqueName: \"kubernetes.io/projected/737d5505-10bc-4e56-b846-01a3af071b38-kube-api-access-j7hk8\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.097094 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-63013308-26d7-4a2a-8b36-252e8653e671\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63013308-26d7-4a2a-8b36-252e8653e671\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.097145 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/737d5505-10bc-4e56-b846-01a3af071b38-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.098775 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/737d5505-10bc-4e56-b846-01a3af071b38-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.099453 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/737d5505-10bc-4e56-b846-01a3af071b38-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.099837 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/737d5505-10bc-4e56-b846-01a3af071b38-config\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.102184 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/737d5505-10bc-4e56-b846-01a3af071b38-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.103031 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/737d5505-10bc-4e56-b846-01a3af071b38-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.117492 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.117547 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-63013308-26d7-4a2a-8b36-252e8653e671\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63013308-26d7-4a2a-8b36-252e8653e671\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/45673400a517db2bba603387b9a8931b6deab901a7561260b1a181375ef1cccd/globalmount\"" pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.121941 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7hk8\" (UniqueName: \"kubernetes.io/projected/737d5505-10bc-4e56-b846-01a3af071b38-kube-api-access-j7hk8\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.127923 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/737d5505-10bc-4e56-b846-01a3af071b38-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.170698 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-63013308-26d7-4a2a-8b36-252e8653e671\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63013308-26d7-4a2a-8b36-252e8653e671\") pod \"ovsdbserver-sb-0\" (UID: \"737d5505-10bc-4e56-b846-01a3af071b38\") " pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:51 crc kubenswrapper[4777]: I1124 17:17:51.206508 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.048322 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.052041 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.054554 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-44n9t" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.054717 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.055391 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.056924 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.069382 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.136489 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4ccca46a-1cac-4693-8490-2612b138c9d5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.136530 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ccca46a-1cac-4693-8490-2612b138c9d5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.136655 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ccca46a-1cac-4693-8490-2612b138c9d5-config\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.136702 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ccca46a-1cac-4693-8490-2612b138c9d5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.136889 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ccca46a-1cac-4693-8490-2612b138c9d5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.137101 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ccca46a-1cac-4693-8490-2612b138c9d5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.137278 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7df03c06-27ae-4cdf-b9e6-933c762d0b44\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7df03c06-27ae-4cdf-b9e6-933c762d0b44\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.137374 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89j6g\" (UniqueName: \"kubernetes.io/projected/4ccca46a-1cac-4693-8490-2612b138c9d5-kube-api-access-89j6g\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.238912 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ccca46a-1cac-4693-8490-2612b138c9d5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.239035 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7df03c06-27ae-4cdf-b9e6-933c762d0b44\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7df03c06-27ae-4cdf-b9e6-933c762d0b44\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.239076 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89j6g\" (UniqueName: \"kubernetes.io/projected/4ccca46a-1cac-4693-8490-2612b138c9d5-kube-api-access-89j6g\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.239165 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4ccca46a-1cac-4693-8490-2612b138c9d5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.239189 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ccca46a-1cac-4693-8490-2612b138c9d5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.239238 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ccca46a-1cac-4693-8490-2612b138c9d5-config\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.239261 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ccca46a-1cac-4693-8490-2612b138c9d5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.239307 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ccca46a-1cac-4693-8490-2612b138c9d5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.240692 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ccca46a-1cac-4693-8490-2612b138c9d5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.240876 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4ccca46a-1cac-4693-8490-2612b138c9d5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.242451 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ccca46a-1cac-4693-8490-2612b138c9d5-config\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.246929 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ccca46a-1cac-4693-8490-2612b138c9d5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.259149 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ccca46a-1cac-4693-8490-2612b138c9d5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.276884 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.276937 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7df03c06-27ae-4cdf-b9e6-933c762d0b44\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7df03c06-27ae-4cdf-b9e6-933c762d0b44\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/456375e5015af9d5f91647ec1234626beeafe6ca91e8ca5e2752654e561f78f5/globalmount\"" pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.278449 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ccca46a-1cac-4693-8490-2612b138c9d5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.290498 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89j6g\" (UniqueName: \"kubernetes.io/projected/4ccca46a-1cac-4693-8490-2612b138c9d5-kube-api-access-89j6g\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.568410 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7df03c06-27ae-4cdf-b9e6-933c762d0b44\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7df03c06-27ae-4cdf-b9e6-933c762d0b44\") pod \"ovsdbserver-nb-0\" (UID: \"4ccca46a-1cac-4693-8490-2612b138c9d5\") " pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:53 crc kubenswrapper[4777]: I1124 17:17:53.683153 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.239331 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8"] Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.240692 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.245365 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-distributor-grpc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.245862 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-ca-bundle" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.246396 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-dockercfg-4mhb5" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.246532 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-distributor-http" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.246744 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-config" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.263393 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8"] Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.308869 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxlnz\" (UniqueName: \"kubernetes.io/projected/9c73a207-d21c-4157-b27a-41656a5f6af1-kube-api-access-qxlnz\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jg9l8\" (UID: \"9c73a207-d21c-4157-b27a-41656a5f6af1\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.308927 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c73a207-d21c-4157-b27a-41656a5f6af1-config\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jg9l8\" (UID: \"9c73a207-d21c-4157-b27a-41656a5f6af1\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.309006 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c73a207-d21c-4157-b27a-41656a5f6af1-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jg9l8\" (UID: \"9c73a207-d21c-4157-b27a-41656a5f6af1\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.309098 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/9c73a207-d21c-4157-b27a-41656a5f6af1-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jg9l8\" (UID: \"9c73a207-d21c-4157-b27a-41656a5f6af1\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.309166 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/9c73a207-d21c-4157-b27a-41656a5f6af1-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jg9l8\" (UID: \"9c73a207-d21c-4157-b27a-41656a5f6af1\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.410245 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-querier-548665d79b-tqjct"] Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.411090 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/9c73a207-d21c-4157-b27a-41656a5f6af1-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jg9l8\" (UID: \"9c73a207-d21c-4157-b27a-41656a5f6af1\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.411161 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxlnz\" (UniqueName: \"kubernetes.io/projected/9c73a207-d21c-4157-b27a-41656a5f6af1-kube-api-access-qxlnz\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jg9l8\" (UID: \"9c73a207-d21c-4157-b27a-41656a5f6af1\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.411183 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c73a207-d21c-4157-b27a-41656a5f6af1-config\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jg9l8\" (UID: \"9c73a207-d21c-4157-b27a-41656a5f6af1\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.411216 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c73a207-d21c-4157-b27a-41656a5f6af1-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jg9l8\" (UID: \"9c73a207-d21c-4157-b27a-41656a5f6af1\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.411283 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/9c73a207-d21c-4157-b27a-41656a5f6af1-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jg9l8\" (UID: \"9c73a207-d21c-4157-b27a-41656a5f6af1\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.413374 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c73a207-d21c-4157-b27a-41656a5f6af1-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jg9l8\" (UID: \"9c73a207-d21c-4157-b27a-41656a5f6af1\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.413529 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c73a207-d21c-4157-b27a-41656a5f6af1-config\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jg9l8\" (UID: \"9c73a207-d21c-4157-b27a-41656a5f6af1\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.420867 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/9c73a207-d21c-4157-b27a-41656a5f6af1-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jg9l8\" (UID: \"9c73a207-d21c-4157-b27a-41656a5f6af1\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.435533 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/9c73a207-d21c-4157-b27a-41656a5f6af1-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jg9l8\" (UID: \"9c73a207-d21c-4157-b27a-41656a5f6af1\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.439036 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.440834 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxlnz\" (UniqueName: \"kubernetes.io/projected/9c73a207-d21c-4157-b27a-41656a5f6af1-kube-api-access-qxlnz\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jg9l8\" (UID: \"9c73a207-d21c-4157-b27a-41656a5f6af1\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.445668 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-querier-548665d79b-tqjct"] Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.452102 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-loki-s3" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.452243 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-querier-http" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.452491 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-querier-grpc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.510431 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc"] Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.511837 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.514794 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-config\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.514854 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvxb7\" (UniqueName: \"kubernetes.io/projected/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-kube-api-access-kvxb7\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.515286 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.515547 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.515595 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.515598 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-query-frontend-http" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.515691 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.519312 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-query-frontend-grpc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.545245 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc"] Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.573048 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.619712 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.619760 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83-config\") pod \"cloudkitty-lokistack-query-frontend-779849886d-x7zhc\" (UID: \"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.619799 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-config\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.619821 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvxb7\" (UniqueName: \"kubernetes.io/projected/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-kube-api-access-kvxb7\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.619838 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-779849886d-x7zhc\" (UID: \"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.619890 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.619908 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7mxs\" (UniqueName: \"kubernetes.io/projected/0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83-kube-api-access-c7mxs\") pod \"cloudkitty-lokistack-query-frontend-779849886d-x7zhc\" (UID: \"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.619931 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-779849886d-x7zhc\" (UID: \"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.619956 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-779849886d-x7zhc\" (UID: \"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.619995 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.620023 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.621110 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.622078 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-config\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.623547 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8"] Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.626500 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.627014 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.628601 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.637579 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-ca" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.637846 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-http" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.638013 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-gateway-ca-bundle" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.638135 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-gateway" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.638264 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-client-http" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.638268 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.646745 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.646921 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvxb7\" (UniqueName: \"kubernetes.io/projected/c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb-kube-api-access-kvxb7\") pod \"cloudkitty-lokistack-querier-548665d79b-tqjct\" (UID: \"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.651495 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh"] Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.653315 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.662004 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-dockercfg-wq88w" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.681021 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8"] Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.703082 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh"] Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.724739 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7mxs\" (UniqueName: \"kubernetes.io/projected/0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83-kube-api-access-c7mxs\") pod \"cloudkitty-lokistack-query-frontend-779849886d-x7zhc\" (UID: \"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.725025 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1aec5e66-69c2-404c-865a-0995f09fd435-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.725050 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghm4t\" (UniqueName: \"kubernetes.io/projected/1aec5e66-69c2-404c-865a-0995f09fd435-kube-api-access-ghm4t\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.725071 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-779849886d-x7zhc\" (UID: \"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.725098 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/1aec5e66-69c2-404c-865a-0995f09fd435-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.725735 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-779849886d-x7zhc\" (UID: \"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.725766 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.750404 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.750530 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1aec5e66-69c2-404c-865a-0995f09fd435-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.750565 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1aec5e66-69c2-404c-865a-0995f09fd435-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.750630 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.750664 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83-config\") pod \"cloudkitty-lokistack-query-frontend-779849886d-x7zhc\" (UID: \"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.750686 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/1aec5e66-69c2-404c-865a-0995f09fd435-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.750724 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.750757 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.750781 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpmrb\" (UniqueName: \"kubernetes.io/projected/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-kube-api-access-zpmrb\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.750802 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/1aec5e66-69c2-404c-865a-0995f09fd435-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.750844 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.750895 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-779849886d-x7zhc\" (UID: \"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.750928 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.750961 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.751069 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/1aec5e66-69c2-404c-865a-0995f09fd435-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.751114 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/1aec5e66-69c2-404c-865a-0995f09fd435-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.726523 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-779849886d-x7zhc\" (UID: \"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.740254 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-779849886d-x7zhc\" (UID: \"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.752666 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7mxs\" (UniqueName: \"kubernetes.io/projected/0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83-kube-api-access-c7mxs\") pod \"cloudkitty-lokistack-query-frontend-779849886d-x7zhc\" (UID: \"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.758886 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-779849886d-x7zhc\" (UID: \"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.761855 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83-config\") pod \"cloudkitty-lokistack-query-frontend-779849886d-x7zhc\" (UID: \"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.803094 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.847657 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853096 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853168 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853222 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1aec5e66-69c2-404c-865a-0995f09fd435-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853245 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1aec5e66-69c2-404c-865a-0995f09fd435-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853299 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853320 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/1aec5e66-69c2-404c-865a-0995f09fd435-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853342 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853361 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853393 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpmrb\" (UniqueName: \"kubernetes.io/projected/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-kube-api-access-zpmrb\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853409 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/1aec5e66-69c2-404c-865a-0995f09fd435-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853432 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853470 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853490 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853537 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/1aec5e66-69c2-404c-865a-0995f09fd435-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853568 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/1aec5e66-69c2-404c-865a-0995f09fd435-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853611 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1aec5e66-69c2-404c-865a-0995f09fd435-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853629 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghm4t\" (UniqueName: \"kubernetes.io/projected/1aec5e66-69c2-404c-865a-0995f09fd435-kube-api-access-ghm4t\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.853647 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/1aec5e66-69c2-404c-865a-0995f09fd435-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.854002 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.854750 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.854786 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.854788 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/1aec5e66-69c2-404c-865a-0995f09fd435-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: E1124 17:17:57.854866 4777 secret.go:188] Couldn't get secret openstack/cloudkitty-lokistack-gateway-http: secret "cloudkitty-lokistack-gateway-http" not found Nov 24 17:17:57 crc kubenswrapper[4777]: E1124 17:17:57.854907 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1aec5e66-69c2-404c-865a-0995f09fd435-tls-secret podName:1aec5e66-69c2-404c-865a-0995f09fd435 nodeName:}" failed. No retries permitted until 2025-11-24 17:17:58.354892924 +0000 UTC m=+1036.514087973 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/1aec5e66-69c2-404c-865a-0995f09fd435-tls-secret") pod "cloudkitty-lokistack-gateway-76cc998948-znzk8" (UID: "1aec5e66-69c2-404c-865a-0995f09fd435") : secret "cloudkitty-lokistack-gateway-http" not found Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.855728 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/1aec5e66-69c2-404c-865a-0995f09fd435-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: E1124 17:17:57.856215 4777 secret.go:188] Couldn't get secret openstack/cloudkitty-lokistack-gateway-http: secret "cloudkitty-lokistack-gateway-http" not found Nov 24 17:17:57 crc kubenswrapper[4777]: E1124 17:17:57.856296 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-tls-secret podName:f1facf54-ce8f-441b-bdef-d864d3bc5dd7 nodeName:}" failed. No retries permitted until 2025-11-24 17:17:58.356276413 +0000 UTC m=+1036.515471462 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-tls-secret") pod "cloudkitty-lokistack-gateway-76cc998948-qskdh" (UID: "f1facf54-ce8f-441b-bdef-d864d3bc5dd7") : secret "cloudkitty-lokistack-gateway-http" not found Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.856626 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.856849 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1aec5e66-69c2-404c-865a-0995f09fd435-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.856856 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/1aec5e66-69c2-404c-865a-0995f09fd435-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.857158 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.857628 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1aec5e66-69c2-404c-865a-0995f09fd435-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.857716 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1aec5e66-69c2-404c-865a-0995f09fd435-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.857886 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.858751 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.862721 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/1aec5e66-69c2-404c-865a-0995f09fd435-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.871301 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghm4t\" (UniqueName: \"kubernetes.io/projected/1aec5e66-69c2-404c-865a-0995f09fd435-kube-api-access-ghm4t\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:57 crc kubenswrapper[4777]: I1124 17:17:57.873323 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpmrb\" (UniqueName: \"kubernetes.io/projected/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-kube-api-access-zpmrb\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.362506 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/1aec5e66-69c2-404c-865a-0995f09fd435-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.362695 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.369268 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/1aec5e66-69c2-404c-865a-0995f09fd435-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-znzk8\" (UID: \"1aec5e66-69c2-404c-865a-0995f09fd435\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.371629 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/f1facf54-ce8f-441b-bdef-d864d3bc5dd7-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-qskdh\" (UID: \"f1facf54-ce8f-441b-bdef-d864d3bc5dd7\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.386272 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.387959 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.391326 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-ingester-http" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.400932 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-ingester-grpc" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.415938 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.464456 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/f68bfbdc-e400-488d-9f32-ea90fd1160e3-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.464882 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.464913 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68bfbdc-e400-488d-9f32-ea90fd1160e3-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.464937 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnhqb\" (UniqueName: \"kubernetes.io/projected/f68bfbdc-e400-488d-9f32-ea90fd1160e3-kube-api-access-xnhqb\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.464962 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f68bfbdc-e400-488d-9f32-ea90fd1160e3-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.465047 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/f68bfbdc-e400-488d-9f32-ea90fd1160e3-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.465070 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.465122 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/f68bfbdc-e400-488d-9f32-ea90fd1160e3-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.495944 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.497409 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.507422 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-compactor-grpc" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.507563 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-compactor-http" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.532462 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.566347 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e99dec76-e21f-4aab-bdc9-ae098391ee6e-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.566435 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/f68bfbdc-e400-488d-9f32-ea90fd1160e3-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.566499 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/e99dec76-e21f-4aab-bdc9-ae098391ee6e-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.566536 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e99dec76-e21f-4aab-bdc9-ae098391ee6e-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.566557 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/e99dec76-e21f-4aab-bdc9-ae098391ee6e-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.566599 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwx4r\" (UniqueName: \"kubernetes.io/projected/e99dec76-e21f-4aab-bdc9-ae098391ee6e-kube-api-access-bwx4r\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.566631 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/f68bfbdc-e400-488d-9f32-ea90fd1160e3-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.566801 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.566869 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68bfbdc-e400-488d-9f32-ea90fd1160e3-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.566900 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.566934 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnhqb\" (UniqueName: \"kubernetes.io/projected/f68bfbdc-e400-488d-9f32-ea90fd1160e3-kube-api-access-xnhqb\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.567015 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f68bfbdc-e400-488d-9f32-ea90fd1160e3-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.567084 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/e99dec76-e21f-4aab-bdc9-ae098391ee6e-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.567156 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/f68bfbdc-e400-488d-9f32-ea90fd1160e3-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.567212 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.568060 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.568218 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.568634 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f68bfbdc-e400-488d-9f32-ea90fd1160e3-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.569397 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68bfbdc-e400-488d-9f32-ea90fd1160e3-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.569750 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/f68bfbdc-e400-488d-9f32-ea90fd1160e3-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.573522 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/f68bfbdc-e400-488d-9f32-ea90fd1160e3-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.574094 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/f68bfbdc-e400-488d-9f32-ea90fd1160e3-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.593194 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.593691 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnhqb\" (UniqueName: \"kubernetes.io/projected/f68bfbdc-e400-488d-9f32-ea90fd1160e3-kube-api-access-xnhqb\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.608914 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.609042 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.609156 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.610591 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.617736 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-index-gateway-grpc" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.618145 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.618206 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-index-gateway-http" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.635725 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"f68bfbdc-e400-488d-9f32-ea90fd1160e3\") " pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.668841 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e99dec76-e21f-4aab-bdc9-ae098391ee6e-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.668902 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/e99dec76-e21f-4aab-bdc9-ae098391ee6e-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.668935 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwx4r\" (UniqueName: \"kubernetes.io/projected/e99dec76-e21f-4aab-bdc9-ae098391ee6e-kube-api-access-bwx4r\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.668995 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.669023 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.669045 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/e99dec76-e21f-4aab-bdc9-ae098391ee6e-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.669066 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.669104 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e99dec76-e21f-4aab-bdc9-ae098391ee6e-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.669130 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.669162 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.669190 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.669208 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j92pc\" (UniqueName: \"kubernetes.io/projected/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-kube-api-access-j92pc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.669276 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/e99dec76-e21f-4aab-bdc9-ae098391ee6e-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.669297 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.670176 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e99dec76-e21f-4aab-bdc9-ae098391ee6e-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.671116 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.671637 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e99dec76-e21f-4aab-bdc9-ae098391ee6e-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.675190 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/e99dec76-e21f-4aab-bdc9-ae098391ee6e-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.675741 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/e99dec76-e21f-4aab-bdc9-ae098391ee6e-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.679774 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/e99dec76-e21f-4aab-bdc9-ae098391ee6e-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.694112 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.697159 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwx4r\" (UniqueName: \"kubernetes.io/projected/e99dec76-e21f-4aab-bdc9-ae098391ee6e-kube-api-access-bwx4r\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"e99dec76-e21f-4aab-bdc9-ae098391ee6e\") " pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.715427 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.770880 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.770928 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.770987 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.771025 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.771055 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.771074 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j92pc\" (UniqueName: \"kubernetes.io/projected/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-kube-api-access-j92pc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.771093 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.771753 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.772635 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.772727 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.775897 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.777453 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.781549 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.789803 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j92pc\" (UniqueName: \"kubernetes.io/projected/8ddeb3e7-8676-40b5-918b-6d8ac12a2b35-kube-api-access-j92pc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.798655 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:17:58 crc kubenswrapper[4777]: I1124 17:17:58.819449 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:17:59 crc kubenswrapper[4777]: I1124 17:17:59.034744 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:18:05 crc kubenswrapper[4777]: E1124 17:18:05.485227 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Nov 24 17:18:05 crc kubenswrapper[4777]: E1124 17:18:05.489189 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n57ch5cch597h5d5h56ch9dhc6h6h5cbh8fh67h5c8h95h7chcfh697h677h95h5ch5fdh679hfbh55dh5bfh84h577h6fh5h699h7bh659h584q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j2xb9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(0254e0e3-f9b9-4017-954a-14d3d21e37c5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:18:05 crc kubenswrapper[4777]: E1124 17:18:05.490400 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="0254e0e3-f9b9-4017-954a-14d3d21e37c5" Nov 24 17:18:06 crc kubenswrapper[4777]: E1124 17:18:06.472992 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="0254e0e3-f9b9-4017-954a-14d3d21e37c5" Nov 24 17:18:44 crc kubenswrapper[4777]: I1124 17:18:44.270484 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:18:44 crc kubenswrapper[4777]: I1124 17:18:44.271223 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:18:57 crc kubenswrapper[4777]: E1124 17:18:57.364651 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Nov 24 17:18:57 crc kubenswrapper[4777]: E1124 17:18:57.365819 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2ltf9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(4c707277-46e4-42ea-9076-566b2d01f544): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:18:57 crc kubenswrapper[4777]: E1124 17:18:57.367162 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="4c707277-46e4-42ea-9076-566b2d01f544" Nov 24 17:18:57 crc kubenswrapper[4777]: E1124 17:18:57.548127 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:d575bb74a591f2d723eb758ea51678520c984b799cb5591b330a41254792e05c: Get \"https://quay.io/v2/auth?account=openshift-release-dev%2Bocm_access_1b89217552bc42d1be3fb06a1aed001a&scope=repository%3Apodified-antelope-centos9%2Fopenstack-neutron-server%3Apull&service=quay.io\": context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 17:18:57 crc kubenswrapper[4777]: E1124 17:18:57.548776 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2ltzh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-j9kwb_openstack(3370ce09-05a5-4d40-b89b-d75720b39677): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:d575bb74a591f2d723eb758ea51678520c984b799cb5591b330a41254792e05c: Get \"https://quay.io/v2/auth?account=openshift-release-dev%2Bocm_access_1b89217552bc42d1be3fb06a1aed001a&scope=repository%3Apodified-antelope-centos9%2Fopenstack-neutron-server%3Apull&service=quay.io\": context canceled" logger="UnhandledError" Nov 24 17:18:57 crc kubenswrapper[4777]: E1124 17:18:57.550256 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:d575bb74a591f2d723eb758ea51678520c984b799cb5591b330a41254792e05c: Get \\\"https://quay.io/v2/auth?account=openshift-release-dev%2Bocm_access_1b89217552bc42d1be3fb06a1aed001a&scope=repository%3Apodified-antelope-centos9%2Fopenstack-neutron-server%3Apull&service=quay.io\\\": context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" podUID="3370ce09-05a5-4d40-b89b-d75720b39677" Nov 24 17:18:57 crc kubenswrapper[4777]: E1124 17:18:57.677630 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Nov 24 17:18:57 crc kubenswrapper[4777]: E1124 17:18:57.677793 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cbf6t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(7163366e-9a39-4379-af0a-d6c7c998ab7f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:18:57 crc kubenswrapper[4777]: E1124 17:18:57.678941 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="7163366e-9a39-4379-af0a-d6c7c998ab7f" Nov 24 17:18:58 crc kubenswrapper[4777]: E1124 17:18:58.031433 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" podUID="3370ce09-05a5-4d40-b89b-d75720b39677" Nov 24 17:18:58 crc kubenswrapper[4777]: E1124 17:18:58.031489 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="4c707277-46e4-42ea-9076-566b2d01f544" Nov 24 17:18:58 crc kubenswrapper[4777]: E1124 17:18:58.032892 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="7163366e-9a39-4379-af0a-d6c7c998ab7f" Nov 24 17:19:00 crc kubenswrapper[4777]: E1124 17:19:00.707410 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Nov 24 17:19:00 crc kubenswrapper[4777]: E1124 17:19:00.707881 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mxwxq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(4eba3f22-b7ad-464d-a0c9-437c449d124b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:19:00 crc kubenswrapper[4777]: E1124 17:19:00.709169 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="4eba3f22-b7ad-464d-a0c9-437c449d124b" Nov 24 17:19:01 crc kubenswrapper[4777]: E1124 17:19:01.520698 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="4eba3f22-b7ad-464d-a0c9-437c449d124b" Nov 24 17:19:01 crc kubenswrapper[4777]: E1124 17:19:01.721614 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Nov 24 17:19:01 crc kubenswrapper[4777]: E1124 17:19:01.721860 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w78cs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(1d71a086-37e0-4409-9db9-66e6f1374406): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:19:01 crc kubenswrapper[4777]: E1124 17:19:01.724061 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="1d71a086-37e0-4409-9db9-66e6f1374406" Nov 24 17:19:02 crc kubenswrapper[4777]: E1124 17:19:02.069201 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="1d71a086-37e0-4409-9db9-66e6f1374406" Nov 24 17:19:05 crc kubenswrapper[4777]: E1124 17:19:05.127745 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 17:19:05 crc kubenswrapper[4777]: E1124 17:19:05.127994 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gjdfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-dx7jp_openstack(180cb346-ec2e-4122-8859-be44e2a33e1e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:19:05 crc kubenswrapper[4777]: E1124 17:19:05.129344 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" podUID="180cb346-ec2e-4122-8859-be44e2a33e1e" Nov 24 17:19:05 crc kubenswrapper[4777]: I1124 17:19:05.732777 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 17:19:05 crc kubenswrapper[4777]: I1124 17:19:05.811432 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-q4l2k"] Nov 24 17:19:05 crc kubenswrapper[4777]: E1124 17:19:05.850167 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 17:19:05 crc kubenswrapper[4777]: E1124 17:19:05.850331 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6zkbz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-zdtpv_openstack(fea7030b-512f-4835-bde8-c9f9905aa91e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:19:05 crc kubenswrapper[4777]: E1124 17:19:05.851602 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-zdtpv" podUID="fea7030b-512f-4835-bde8-c9f9905aa91e" Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.060138 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:19:06 crc kubenswrapper[4777]: W1124 17:19:06.060699 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod155b4504_c156_4af8_a9e0_1dc66c9d7d43.slice/crio-a97fdefbd5723c9e2d7d6333a32e3b2b7926b259d827a2b1b9e38959f41722bd WatchSource:0}: Error finding container a97fdefbd5723c9e2d7d6333a32e3b2b7926b259d827a2b1b9e38959f41722bd: Status 404 returned error can't find the container with id a97fdefbd5723c9e2d7d6333a32e3b2b7926b259d827a2b1b9e38959f41722bd Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.089955 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 17:19:06 crc kubenswrapper[4777]: W1124 17:19:06.091975 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9ab961d_98f3_4a5e_9583_a168fa95f7b4.slice/crio-2771454974f38c8e15ef848abdd808509a7d82973af1197fff72ba4575e319c8 WatchSource:0}: Error finding container 2771454974f38c8e15ef848abdd808509a7d82973af1197fff72ba4575e319c8: Status 404 returned error can't find the container with id 2771454974f38c8e15ef848abdd808509a7d82973af1197fff72ba4575e319c8 Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.123165 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"155b4504-c156-4af8-a9e0-1dc66c9d7d43","Type":"ContainerStarted","Data":"a97fdefbd5723c9e2d7d6333a32e3b2b7926b259d827a2b1b9e38959f41722bd"} Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.126205 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f9ab961d-98f3-4a5e-9583-a168fa95f7b4","Type":"ContainerStarted","Data":"2771454974f38c8e15ef848abdd808509a7d82973af1197fff72ba4575e319c8"} Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.127421 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-q4l2k" event={"ID":"73928e73-667c-4b69-aed9-72f32012fbdc","Type":"ContainerStarted","Data":"183a9d1b3408c369534341df4c718b044d592f39a1f8c16524b6a0d4e5e360d5"} Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.130476 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"737d5505-10bc-4e56-b846-01a3af071b38","Type":"ContainerStarted","Data":"630a6630e83f49b27ce7917107e12292c35a543835b68e443bc0029b1d167535"} Nov 24 17:19:06 crc kubenswrapper[4777]: E1124 17:19:06.132320 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" podUID="180cb346-ec2e-4122-8859-be44e2a33e1e" Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.151680 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-dkbds"] Nov 24 17:19:06 crc kubenswrapper[4777]: W1124 17:19:06.171720 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3641f728_c179_4f10_93c2_98f74763e77f.slice/crio-efd246171ae3a87c135ddd4003499ba23ed9648872e9dcb2232d9db993287852 WatchSource:0}: Error finding container efd246171ae3a87c135ddd4003499ba23ed9648872e9dcb2232d9db993287852: Status 404 returned error can't find the container with id efd246171ae3a87c135ddd4003499ba23ed9648872e9dcb2232d9db993287852 Nov 24 17:19:06 crc kubenswrapper[4777]: E1124 17:19:06.246086 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 17:19:06 crc kubenswrapper[4777]: E1124 17:19:06.246459 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6zkcf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-v2wg2_openstack(919dff3b-4077-4510-b4d0-d1b0080a8357): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:19:06 crc kubenswrapper[4777]: E1124 17:19:06.248347 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" podUID="919dff3b-4077-4510-b4d0-d1b0080a8357" Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.540831 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.588224 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh"] Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.602436 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.609502 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-zdtpv" Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.612433 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8"] Nov 24 17:19:06 crc kubenswrapper[4777]: W1124 17:19:06.613239 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1facf54_ce8f_441b_bdef_d864d3bc5dd7.slice/crio-f0c5945a1749663992922adee6142c345bfa047f98c24d46093a410327116ae5 WatchSource:0}: Error finding container f0c5945a1749663992922adee6142c345bfa047f98c24d46093a410327116ae5: Status 404 returned error can't find the container with id f0c5945a1749663992922adee6142c345bfa047f98c24d46093a410327116ae5 Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.628294 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8"] Nov 24 17:19:06 crc kubenswrapper[4777]: W1124 17:19:06.628480 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d98d291_cc6c_4c5b_90a8_fa2bca9e9f83.slice/crio-df543102c176d76fb060adb9caa155749de3976775bacb86fcbe63d12036fc6e WatchSource:0}: Error finding container df543102c176d76fb060adb9caa155749de3976775bacb86fcbe63d12036fc6e: Status 404 returned error can't find the container with id df543102c176d76fb060adb9caa155749de3976775bacb86fcbe63d12036fc6e Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.635090 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc"] Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.641837 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.650108 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.683178 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 17:19:06 crc kubenswrapper[4777]: W1124 17:19:06.683870 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ccca46a_1cac_4693_8490_2612b138c9d5.slice/crio-078591a38abfa9c6def7c8299542e01c7421a781476d55c143ab42622f645677 WatchSource:0}: Error finding container 078591a38abfa9c6def7c8299542e01c7421a781476d55c143ab42622f645677: Status 404 returned error can't find the container with id 078591a38abfa9c6def7c8299542e01c7421a781476d55c143ab42622f645677 Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.764521 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zkbz\" (UniqueName: \"kubernetes.io/projected/fea7030b-512f-4835-bde8-c9f9905aa91e-kube-api-access-6zkbz\") pod \"fea7030b-512f-4835-bde8-c9f9905aa91e\" (UID: \"fea7030b-512f-4835-bde8-c9f9905aa91e\") " Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.764566 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fea7030b-512f-4835-bde8-c9f9905aa91e-config\") pod \"fea7030b-512f-4835-bde8-c9f9905aa91e\" (UID: \"fea7030b-512f-4835-bde8-c9f9905aa91e\") " Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.765136 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fea7030b-512f-4835-bde8-c9f9905aa91e-config" (OuterVolumeSpecName: "config") pod "fea7030b-512f-4835-bde8-c9f9905aa91e" (UID: "fea7030b-512f-4835-bde8-c9f9905aa91e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.775755 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fea7030b-512f-4835-bde8-c9f9905aa91e-kube-api-access-6zkbz" (OuterVolumeSpecName: "kube-api-access-6zkbz") pod "fea7030b-512f-4835-bde8-c9f9905aa91e" (UID: "fea7030b-512f-4835-bde8-c9f9905aa91e"). InnerVolumeSpecName "kube-api-access-6zkbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.779017 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-querier-548665d79b-tqjct"] Nov 24 17:19:06 crc kubenswrapper[4777]: W1124 17:19:06.796801 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0995ddd_f0a2_4137_ac83_2aaae1ce1aeb.slice/crio-6daa95a42c91d91e75e6663eb62da2ce56a92994bf111de51a29be766186e5e8 WatchSource:0}: Error finding container 6daa95a42c91d91e75e6663eb62da2ce56a92994bf111de51a29be766186e5e8: Status 404 returned error can't find the container with id 6daa95a42c91d91e75e6663eb62da2ce56a92994bf111de51a29be766186e5e8 Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.866443 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zkbz\" (UniqueName: \"kubernetes.io/projected/fea7030b-512f-4835-bde8-c9f9905aa91e-kube-api-access-6zkbz\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:06 crc kubenswrapper[4777]: I1124 17:19:06.866525 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fea7030b-512f-4835-bde8-c9f9905aa91e-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.139828 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" event={"ID":"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83","Type":"ContainerStarted","Data":"df543102c176d76fb060adb9caa155749de3976775bacb86fcbe63d12036fc6e"} Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.141472 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"055503f1-ab1c-47db-be36-37a3f572be32","Type":"ContainerStarted","Data":"e361b65252426334dc982c254046ea1b7b2c834b7d59170fc4efc404e80a34b0"} Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.142876 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-index-gateway-0" event={"ID":"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35","Type":"ContainerStarted","Data":"27ccb1769a6cd3592400818d3e4408e1830f8de72bb34fd31f8c49cc5c374b8b"} Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.144292 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-ingester-0" event={"ID":"f68bfbdc-e400-488d-9f32-ea90fd1160e3","Type":"ContainerStarted","Data":"da14981ebf4dad6eb0729525998663c9acc5bba7528db6971bb916e853d09ad2"} Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.145349 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" event={"ID":"f1facf54-ce8f-441b-bdef-d864d3bc5dd7","Type":"ContainerStarted","Data":"f0c5945a1749663992922adee6142c345bfa047f98c24d46093a410327116ae5"} Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.146359 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" event={"ID":"1aec5e66-69c2-404c-865a-0995f09fd435","Type":"ContainerStarted","Data":"ada97ae04ed4767221748ac5671d94b2d0a85ca0b98612ee8a05c6c1de373c02"} Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.147988 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-zdtpv" event={"ID":"fea7030b-512f-4835-bde8-c9f9905aa91e","Type":"ContainerDied","Data":"e41d41c321ea45325b88d2471a3dd8692a3f816c7ff414ece59bb088e1a622eb"} Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.148051 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-zdtpv" Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.149747 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-compactor-0" event={"ID":"e99dec76-e21f-4aab-bdc9-ae098391ee6e","Type":"ContainerStarted","Data":"e9fb00ee74fac1e7c8587387d5a5e9bc164bdc7ef35c9b4d2bcadf4ff3177ba3"} Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.152568 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0254e0e3-f9b9-4017-954a-14d3d21e37c5","Type":"ContainerStarted","Data":"630763b052cd6447c5b4e6d5ab5b0a97194a648af685c2c3999a98e617c3484a"} Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.152789 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.154451 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" event={"ID":"9c73a207-d21c-4157-b27a-41656a5f6af1","Type":"ContainerStarted","Data":"af53ec51c4b4826b6aa97d9f76ce6d23a5e0764c629e363d037c2be5ba0614ef"} Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.156248 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" event={"ID":"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb","Type":"ContainerStarted","Data":"6daa95a42c91d91e75e6663eb62da2ce56a92994bf111de51a29be766186e5e8"} Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.158939 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dkbds" event={"ID":"3641f728-c179-4f10-93c2-98f74763e77f","Type":"ContainerStarted","Data":"efd246171ae3a87c135ddd4003499ba23ed9648872e9dcb2232d9db993287852"} Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.159956 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4ccca46a-1cac-4693-8490-2612b138c9d5","Type":"ContainerStarted","Data":"078591a38abfa9c6def7c8299542e01c7421a781476d55c143ab42622f645677"} Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.207500 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=3.010045809 podStartE2EDuration="1m23.207476273s" podCreationTimestamp="2025-11-24 17:17:44 +0000 UTC" firstStartedPulling="2025-11-24 17:17:45.444360849 +0000 UTC m=+1023.603555888" lastFinishedPulling="2025-11-24 17:19:05.641791303 +0000 UTC m=+1103.800986352" observedRunningTime="2025-11-24 17:19:07.173358694 +0000 UTC m=+1105.332553823" watchObservedRunningTime="2025-11-24 17:19:07.207476273 +0000 UTC m=+1105.366671322" Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.273168 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-zdtpv"] Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.295711 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-zdtpv"] Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.549955 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.680710 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/919dff3b-4077-4510-b4d0-d1b0080a8357-config\") pod \"919dff3b-4077-4510-b4d0-d1b0080a8357\" (UID: \"919dff3b-4077-4510-b4d0-d1b0080a8357\") " Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.680782 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/919dff3b-4077-4510-b4d0-d1b0080a8357-dns-svc\") pod \"919dff3b-4077-4510-b4d0-d1b0080a8357\" (UID: \"919dff3b-4077-4510-b4d0-d1b0080a8357\") " Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.680927 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zkcf\" (UniqueName: \"kubernetes.io/projected/919dff3b-4077-4510-b4d0-d1b0080a8357-kube-api-access-6zkcf\") pod \"919dff3b-4077-4510-b4d0-d1b0080a8357\" (UID: \"919dff3b-4077-4510-b4d0-d1b0080a8357\") " Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.681188 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/919dff3b-4077-4510-b4d0-d1b0080a8357-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "919dff3b-4077-4510-b4d0-d1b0080a8357" (UID: "919dff3b-4077-4510-b4d0-d1b0080a8357"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.681231 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/919dff3b-4077-4510-b4d0-d1b0080a8357-config" (OuterVolumeSpecName: "config") pod "919dff3b-4077-4510-b4d0-d1b0080a8357" (UID: "919dff3b-4077-4510-b4d0-d1b0080a8357"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.681707 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/919dff3b-4077-4510-b4d0-d1b0080a8357-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.681725 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/919dff3b-4077-4510-b4d0-d1b0080a8357-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.697954 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/919dff3b-4077-4510-b4d0-d1b0080a8357-kube-api-access-6zkcf" (OuterVolumeSpecName: "kube-api-access-6zkcf") pod "919dff3b-4077-4510-b4d0-d1b0080a8357" (UID: "919dff3b-4077-4510-b4d0-d1b0080a8357"). InnerVolumeSpecName "kube-api-access-6zkcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:19:07 crc kubenswrapper[4777]: I1124 17:19:07.783245 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zkcf\" (UniqueName: \"kubernetes.io/projected/919dff3b-4077-4510-b4d0-d1b0080a8357-kube-api-access-6zkcf\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:08 crc kubenswrapper[4777]: I1124 17:19:08.171534 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" event={"ID":"919dff3b-4077-4510-b4d0-d1b0080a8357","Type":"ContainerDied","Data":"8e56864db48f250282015eeab268622ec6c69d1e46d6594b4c2667c976d82898"} Nov 24 17:19:08 crc kubenswrapper[4777]: I1124 17:19:08.171570 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-v2wg2" Nov 24 17:19:08 crc kubenswrapper[4777]: I1124 17:19:08.261993 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-v2wg2"] Nov 24 17:19:08 crc kubenswrapper[4777]: I1124 17:19:08.262235 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-v2wg2"] Nov 24 17:19:09 crc kubenswrapper[4777]: I1124 17:19:09.261522 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="919dff3b-4077-4510-b4d0-d1b0080a8357" path="/var/lib/kubelet/pods/919dff3b-4077-4510-b4d0-d1b0080a8357/volumes" Nov 24 17:19:09 crc kubenswrapper[4777]: I1124 17:19:09.261923 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fea7030b-512f-4835-bde8-c9f9905aa91e" path="/var/lib/kubelet/pods/fea7030b-512f-4835-bde8-c9f9905aa91e/volumes" Nov 24 17:19:14 crc kubenswrapper[4777]: I1124 17:19:14.270449 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:19:14 crc kubenswrapper[4777]: I1124 17:19:14.270915 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:19:14 crc kubenswrapper[4777]: I1124 17:19:14.903252 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 24 17:19:16 crc kubenswrapper[4777]: I1124 17:19:16.539918 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dx7jp"] Nov 24 17:19:16 crc kubenswrapper[4777]: I1124 17:19:16.577309 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-n7hxb"] Nov 24 17:19:16 crc kubenswrapper[4777]: I1124 17:19:16.579420 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" Nov 24 17:19:16 crc kubenswrapper[4777]: I1124 17:19:16.613681 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-n7hxb"] Nov 24 17:19:16 crc kubenswrapper[4777]: I1124 17:19:16.681007 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54920d72-c01f-4d69-bc52-873f086d51c5-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-n7hxb\" (UID: \"54920d72-c01f-4d69-bc52-873f086d51c5\") " pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" Nov 24 17:19:16 crc kubenswrapper[4777]: I1124 17:19:16.681393 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n45n\" (UniqueName: \"kubernetes.io/projected/54920d72-c01f-4d69-bc52-873f086d51c5-kube-api-access-2n45n\") pod \"dnsmasq-dns-7cb5889db5-n7hxb\" (UID: \"54920d72-c01f-4d69-bc52-873f086d51c5\") " pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" Nov 24 17:19:16 crc kubenswrapper[4777]: I1124 17:19:16.681448 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54920d72-c01f-4d69-bc52-873f086d51c5-config\") pod \"dnsmasq-dns-7cb5889db5-n7hxb\" (UID: \"54920d72-c01f-4d69-bc52-873f086d51c5\") " pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" Nov 24 17:19:16 crc kubenswrapper[4777]: I1124 17:19:16.785412 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54920d72-c01f-4d69-bc52-873f086d51c5-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-n7hxb\" (UID: \"54920d72-c01f-4d69-bc52-873f086d51c5\") " pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" Nov 24 17:19:16 crc kubenswrapper[4777]: I1124 17:19:16.785494 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n45n\" (UniqueName: \"kubernetes.io/projected/54920d72-c01f-4d69-bc52-873f086d51c5-kube-api-access-2n45n\") pod \"dnsmasq-dns-7cb5889db5-n7hxb\" (UID: \"54920d72-c01f-4d69-bc52-873f086d51c5\") " pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" Nov 24 17:19:16 crc kubenswrapper[4777]: I1124 17:19:16.786552 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54920d72-c01f-4d69-bc52-873f086d51c5-config\") pod \"dnsmasq-dns-7cb5889db5-n7hxb\" (UID: \"54920d72-c01f-4d69-bc52-873f086d51c5\") " pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" Nov 24 17:19:16 crc kubenswrapper[4777]: I1124 17:19:16.786831 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54920d72-c01f-4d69-bc52-873f086d51c5-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-n7hxb\" (UID: \"54920d72-c01f-4d69-bc52-873f086d51c5\") " pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" Nov 24 17:19:16 crc kubenswrapper[4777]: I1124 17:19:16.787454 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54920d72-c01f-4d69-bc52-873f086d51c5-config\") pod \"dnsmasq-dns-7cb5889db5-n7hxb\" (UID: \"54920d72-c01f-4d69-bc52-873f086d51c5\") " pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" Nov 24 17:19:16 crc kubenswrapper[4777]: I1124 17:19:16.818236 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n45n\" (UniqueName: \"kubernetes.io/projected/54920d72-c01f-4d69-bc52-873f086d51c5-kube-api-access-2n45n\") pod \"dnsmasq-dns-7cb5889db5-n7hxb\" (UID: \"54920d72-c01f-4d69-bc52-873f086d51c5\") " pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" Nov 24 17:19:16 crc kubenswrapper[4777]: I1124 17:19:16.907930 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.646513 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.725622 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.730561 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.730897 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.731104 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-6vglg" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.731107 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.731309 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.831580 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-58047947-f9b3-4683-a8c4-78b5eb6c00f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58047947-f9b3-4683-a8c4-78b5eb6c00f0\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.831877 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9c412d58-6198-4254-932d-5eb90420f8ba-lock\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.832016 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78ww4\" (UniqueName: \"kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-kube-api-access-78ww4\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.832183 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.832340 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9c412d58-6198-4254-932d-5eb90420f8ba-cache\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.933784 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.934149 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9c412d58-6198-4254-932d-5eb90420f8ba-cache\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:17 crc kubenswrapper[4777]: E1124 17:19:17.934052 4777 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 17:19:17 crc kubenswrapper[4777]: E1124 17:19:17.934432 4777 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 17:19:17 crc kubenswrapper[4777]: E1124 17:19:17.934487 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift podName:9c412d58-6198-4254-932d-5eb90420f8ba nodeName:}" failed. No retries permitted until 2025-11-24 17:19:18.434466388 +0000 UTC m=+1116.593661437 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift") pod "swift-storage-0" (UID: "9c412d58-6198-4254-932d-5eb90420f8ba") : configmap "swift-ring-files" not found Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.934702 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9c412d58-6198-4254-932d-5eb90420f8ba-cache\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.934796 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-58047947-f9b3-4683-a8c4-78b5eb6c00f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58047947-f9b3-4683-a8c4-78b5eb6c00f0\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.934948 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9c412d58-6198-4254-932d-5eb90420f8ba-lock\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.935177 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78ww4\" (UniqueName: \"kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-kube-api-access-78ww4\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.935260 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9c412d58-6198-4254-932d-5eb90420f8ba-lock\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.940031 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.940057 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-58047947-f9b3-4683-a8c4-78b5eb6c00f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58047947-f9b3-4683-a8c4-78b5eb6c00f0\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/91a4d8836f38b8d8528414def6b8f4de216480c0cb0d25d51a307c66739c5fac/globalmount\"" pod="openstack/swift-storage-0" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.955853 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78ww4\" (UniqueName: \"kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-kube-api-access-78ww4\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:17 crc kubenswrapper[4777]: I1124 17:19:17.987153 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-58047947-f9b3-4683-a8c4-78b5eb6c00f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58047947-f9b3-4683-a8c4-78b5eb6c00f0\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.226593 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-ss9hw"] Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.227738 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.233231 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.233320 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.233389 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.240937 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-ss9hw"] Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.286382 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-xgm6s"] Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.287826 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.293178 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-ss9hw"] Nov 24 17:19:18 crc kubenswrapper[4777]: E1124 17:19:18.293827 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-sb25s ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-ss9hw" podUID="490542bb-6f1c-4a79-bdca-b6be4ef050cb" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.303960 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-xgm6s"] Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.341220 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-dispersionconf\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.342144 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-swiftconf\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.343132 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb25s\" (UniqueName: \"kubernetes.io/projected/490542bb-6f1c-4a79-bdca-b6be4ef050cb-kube-api-access-sb25s\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.343180 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/490542bb-6f1c-4a79-bdca-b6be4ef050cb-ring-data-devices\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.343226 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/490542bb-6f1c-4a79-bdca-b6be4ef050cb-scripts\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.343336 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/490542bb-6f1c-4a79-bdca-b6be4ef050cb-etc-swift\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.343386 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-combined-ca-bundle\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.444787 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75kwr\" (UniqueName: \"kubernetes.io/projected/3ec24d0c-556b-4a24-8f42-38fed08efe6b-kube-api-access-75kwr\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.444860 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-combined-ca-bundle\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.445013 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ec24d0c-556b-4a24-8f42-38fed08efe6b-etc-swift\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.445068 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-swiftconf\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.445129 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.445166 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-dispersionconf\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.445234 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-swiftconf\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.445283 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-dispersionconf\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: E1124 17:19:18.445328 4777 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 17:19:18 crc kubenswrapper[4777]: E1124 17:19:18.445362 4777 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 17:19:18 crc kubenswrapper[4777]: E1124 17:19:18.445409 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift podName:9c412d58-6198-4254-932d-5eb90420f8ba nodeName:}" failed. No retries permitted until 2025-11-24 17:19:19.44539209 +0000 UTC m=+1117.604587139 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift") pod "swift-storage-0" (UID: "9c412d58-6198-4254-932d-5eb90420f8ba") : configmap "swift-ring-files" not found Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.445336 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb25s\" (UniqueName: \"kubernetes.io/projected/490542bb-6f1c-4a79-bdca-b6be4ef050cb-kube-api-access-sb25s\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.445539 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/490542bb-6f1c-4a79-bdca-b6be4ef050cb-ring-data-devices\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.445595 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/490542bb-6f1c-4a79-bdca-b6be4ef050cb-scripts\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.445703 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ec24d0c-556b-4a24-8f42-38fed08efe6b-scripts\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.445772 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/490542bb-6f1c-4a79-bdca-b6be4ef050cb-etc-swift\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.445861 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-combined-ca-bundle\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.445942 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ec24d0c-556b-4a24-8f42-38fed08efe6b-ring-data-devices\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.446292 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/490542bb-6f1c-4a79-bdca-b6be4ef050cb-etc-swift\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.446410 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/490542bb-6f1c-4a79-bdca-b6be4ef050cb-ring-data-devices\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.446695 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/490542bb-6f1c-4a79-bdca-b6be4ef050cb-scripts\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.454016 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-swiftconf\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.454342 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-dispersionconf\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.461397 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-combined-ca-bundle\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.462822 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb25s\" (UniqueName: \"kubernetes.io/projected/490542bb-6f1c-4a79-bdca-b6be4ef050cb-kube-api-access-sb25s\") pod \"swift-ring-rebalance-ss9hw\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.547263 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ec24d0c-556b-4a24-8f42-38fed08efe6b-scripts\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.547360 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ec24d0c-556b-4a24-8f42-38fed08efe6b-ring-data-devices\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.547424 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75kwr\" (UniqueName: \"kubernetes.io/projected/3ec24d0c-556b-4a24-8f42-38fed08efe6b-kube-api-access-75kwr\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.547449 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-combined-ca-bundle\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.547482 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ec24d0c-556b-4a24-8f42-38fed08efe6b-etc-swift\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.547515 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-swiftconf\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.547617 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-dispersionconf\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.548588 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ec24d0c-556b-4a24-8f42-38fed08efe6b-ring-data-devices\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.548916 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ec24d0c-556b-4a24-8f42-38fed08efe6b-scripts\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.549725 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ec24d0c-556b-4a24-8f42-38fed08efe6b-etc-swift\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.554599 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-combined-ca-bundle\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.554774 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-swiftconf\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.558709 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-dispersionconf\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.572239 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.585325 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75kwr\" (UniqueName: \"kubernetes.io/projected/3ec24d0c-556b-4a24-8f42-38fed08efe6b-kube-api-access-75kwr\") pod \"swift-ring-rebalance-xgm6s\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.601509 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.751766 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/180cb346-ec2e-4122-8859-be44e2a33e1e-config\") pod \"180cb346-ec2e-4122-8859-be44e2a33e1e\" (UID: \"180cb346-ec2e-4122-8859-be44e2a33e1e\") " Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.751884 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjdfq\" (UniqueName: \"kubernetes.io/projected/180cb346-ec2e-4122-8859-be44e2a33e1e-kube-api-access-gjdfq\") pod \"180cb346-ec2e-4122-8859-be44e2a33e1e\" (UID: \"180cb346-ec2e-4122-8859-be44e2a33e1e\") " Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.752021 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/180cb346-ec2e-4122-8859-be44e2a33e1e-dns-svc\") pod \"180cb346-ec2e-4122-8859-be44e2a33e1e\" (UID: \"180cb346-ec2e-4122-8859-be44e2a33e1e\") " Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.752371 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/180cb346-ec2e-4122-8859-be44e2a33e1e-config" (OuterVolumeSpecName: "config") pod "180cb346-ec2e-4122-8859-be44e2a33e1e" (UID: "180cb346-ec2e-4122-8859-be44e2a33e1e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.752449 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/180cb346-ec2e-4122-8859-be44e2a33e1e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "180cb346-ec2e-4122-8859-be44e2a33e1e" (UID: "180cb346-ec2e-4122-8859-be44e2a33e1e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.752550 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/180cb346-ec2e-4122-8859-be44e2a33e1e-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.755055 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/180cb346-ec2e-4122-8859-be44e2a33e1e-kube-api-access-gjdfq" (OuterVolumeSpecName: "kube-api-access-gjdfq") pod "180cb346-ec2e-4122-8859-be44e2a33e1e" (UID: "180cb346-ec2e-4122-8859-be44e2a33e1e"). InnerVolumeSpecName "kube-api-access-gjdfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.854786 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjdfq\" (UniqueName: \"kubernetes.io/projected/180cb346-ec2e-4122-8859-be44e2a33e1e-kube-api-access-gjdfq\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:18 crc kubenswrapper[4777]: I1124 17:19:18.854820 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/180cb346-ec2e-4122-8859-be44e2a33e1e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:19 crc kubenswrapper[4777]: E1124 17:19:19.180065 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified" Nov 24 17:19:19 crc kubenswrapper[4777]: E1124 17:19:19.180646 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-sb,Image:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n59fh75h8fh5d5h64dh644h575h5cfh688h8h57bh67fh5fh567h67h669h564hd6h56ch59dhf9h677h5b6h86h5f4h8dh556h86hch566h547h67cq,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-sb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j7hk8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(737d5505-10bc-4e56-b846-01a3af071b38): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.272932 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.272956 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" event={"ID":"180cb346-ec2e-4122-8859-be44e2a33e1e","Type":"ContainerDied","Data":"d234923475ff0bcc644785693bbe8e63cf649038c0de6dcbf29b34bd614da678"} Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.272939 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dx7jp" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.300226 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.345452 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dx7jp"] Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.359631 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dx7jp"] Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.464102 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-swiftconf\") pod \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.464217 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb25s\" (UniqueName: \"kubernetes.io/projected/490542bb-6f1c-4a79-bdca-b6be4ef050cb-kube-api-access-sb25s\") pod \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.464275 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/490542bb-6f1c-4a79-bdca-b6be4ef050cb-scripts\") pod \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.464317 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/490542bb-6f1c-4a79-bdca-b6be4ef050cb-etc-swift\") pod \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.464348 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-combined-ca-bundle\") pod \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.464424 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-dispersionconf\") pod \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.464578 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/490542bb-6f1c-4a79-bdca-b6be4ef050cb-ring-data-devices\") pod \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\" (UID: \"490542bb-6f1c-4a79-bdca-b6be4ef050cb\") " Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.464911 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:19 crc kubenswrapper[4777]: E1124 17:19:19.465228 4777 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 17:19:19 crc kubenswrapper[4777]: E1124 17:19:19.465244 4777 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 17:19:19 crc kubenswrapper[4777]: E1124 17:19:19.465289 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift podName:9c412d58-6198-4254-932d-5eb90420f8ba nodeName:}" failed. No retries permitted until 2025-11-24 17:19:21.465271846 +0000 UTC m=+1119.624466905 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift") pod "swift-storage-0" (UID: "9c412d58-6198-4254-932d-5eb90420f8ba") : configmap "swift-ring-files" not found Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.466023 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/490542bb-6f1c-4a79-bdca-b6be4ef050cb-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "490542bb-6f1c-4a79-bdca-b6be4ef050cb" (UID: "490542bb-6f1c-4a79-bdca-b6be4ef050cb"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.466205 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/490542bb-6f1c-4a79-bdca-b6be4ef050cb-scripts" (OuterVolumeSpecName: "scripts") pod "490542bb-6f1c-4a79-bdca-b6be4ef050cb" (UID: "490542bb-6f1c-4a79-bdca-b6be4ef050cb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.467625 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/490542bb-6f1c-4a79-bdca-b6be4ef050cb-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "490542bb-6f1c-4a79-bdca-b6be4ef050cb" (UID: "490542bb-6f1c-4a79-bdca-b6be4ef050cb"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.468299 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/490542bb-6f1c-4a79-bdca-b6be4ef050cb-kube-api-access-sb25s" (OuterVolumeSpecName: "kube-api-access-sb25s") pod "490542bb-6f1c-4a79-bdca-b6be4ef050cb" (UID: "490542bb-6f1c-4a79-bdca-b6be4ef050cb"). InnerVolumeSpecName "kube-api-access-sb25s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.468928 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "490542bb-6f1c-4a79-bdca-b6be4ef050cb" (UID: "490542bb-6f1c-4a79-bdca-b6be4ef050cb"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.469096 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "490542bb-6f1c-4a79-bdca-b6be4ef050cb" (UID: "490542bb-6f1c-4a79-bdca-b6be4ef050cb"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.470206 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "490542bb-6f1c-4a79-bdca-b6be4ef050cb" (UID: "490542bb-6f1c-4a79-bdca-b6be4ef050cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.566439 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb25s\" (UniqueName: \"kubernetes.io/projected/490542bb-6f1c-4a79-bdca-b6be4ef050cb-kube-api-access-sb25s\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.566474 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/490542bb-6f1c-4a79-bdca-b6be4ef050cb-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.566484 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.566495 4777 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/490542bb-6f1c-4a79-bdca-b6be4ef050cb-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.566503 4777 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.566513 4777 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/490542bb-6f1c-4a79-bdca-b6be4ef050cb-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:19 crc kubenswrapper[4777]: I1124 17:19:19.566521 4777 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/490542bb-6f1c-4a79-bdca-b6be4ef050cb-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:20 crc kubenswrapper[4777]: I1124 17:19:20.280028 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-ss9hw" Nov 24 17:19:20 crc kubenswrapper[4777]: I1124 17:19:20.346952 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-ss9hw"] Nov 24 17:19:20 crc kubenswrapper[4777]: I1124 17:19:20.352484 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-ss9hw"] Nov 24 17:19:20 crc kubenswrapper[4777]: I1124 17:19:20.933955 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-n7hxb"] Nov 24 17:19:20 crc kubenswrapper[4777]: W1124 17:19:20.964711 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54920d72_c01f_4d69_bc52_873f086d51c5.slice/crio-8c6184ef938e3c4c8cd1a856d68d0c35cacd0bde0e24829cfb1d971647bdc727 WatchSource:0}: Error finding container 8c6184ef938e3c4c8cd1a856d68d0c35cacd0bde0e24829cfb1d971647bdc727: Status 404 returned error can't find the container with id 8c6184ef938e3c4c8cd1a856d68d0c35cacd0bde0e24829cfb1d971647bdc727 Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.053169 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-xgm6s"] Nov 24 17:19:21 crc kubenswrapper[4777]: W1124 17:19:21.151653 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ec24d0c_556b_4a24_8f42_38fed08efe6b.slice/crio-6a410e1deeef1f929bf9dbcfec49f6ad8859a7ced740439460ddfb36e67b0784 WatchSource:0}: Error finding container 6a410e1deeef1f929bf9dbcfec49f6ad8859a7ced740439460ddfb36e67b0784: Status 404 returned error can't find the container with id 6a410e1deeef1f929bf9dbcfec49f6ad8859a7ced740439460ddfb36e67b0784 Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.256200 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="180cb346-ec2e-4122-8859-be44e2a33e1e" path="/var/lib/kubelet/pods/180cb346-ec2e-4122-8859-be44e2a33e1e/volumes" Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.257071 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="490542bb-6f1c-4a79-bdca-b6be4ef050cb" path="/var/lib/kubelet/pods/490542bb-6f1c-4a79-bdca-b6be4ef050cb/volumes" Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.291873 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-index-gateway-0" event={"ID":"8ddeb3e7-8676-40b5-918b-6d8ac12a2b35","Type":"ContainerStarted","Data":"662d1e98da0a6b881ca995f05b36e255cdf8c8fe8e1401281182d1cee9c0fd54"} Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.291961 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.296214 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" event={"ID":"54920d72-c01f-4d69-bc52-873f086d51c5","Type":"ContainerStarted","Data":"8c6184ef938e3c4c8cd1a856d68d0c35cacd0bde0e24829cfb1d971647bdc727"} Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.300528 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" event={"ID":"f1facf54-ce8f-441b-bdef-d864d3bc5dd7","Type":"ContainerStarted","Data":"ac2802e6c24c179e1d0318b551cca0c75c0158cae5eba81af889f451695112e3"} Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.300701 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.302030 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xgm6s" event={"ID":"3ec24d0c-556b-4a24-8f42-38fed08efe6b","Type":"ContainerStarted","Data":"6a410e1deeef1f929bf9dbcfec49f6ad8859a7ced740439460ddfb36e67b0784"} Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.308556 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" event={"ID":"0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83","Type":"ContainerStarted","Data":"1378b5c86c86de05e973741af652c4e043c793a379a94894b42b88249548544f"} Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.309179 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.321178 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.355335 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-index-gateway-0" podStartSLOduration=71.52635406 podStartE2EDuration="1m24.355307933s" podCreationTimestamp="2025-11-24 17:17:57 +0000 UTC" firstStartedPulling="2025-11-24 17:19:06.644115236 +0000 UTC m=+1104.803310285" lastFinishedPulling="2025-11-24 17:19:19.473069079 +0000 UTC m=+1117.632264158" observedRunningTime="2025-11-24 17:19:21.314188013 +0000 UTC m=+1119.473383062" watchObservedRunningTime="2025-11-24 17:19:21.355307933 +0000 UTC m=+1119.514502982" Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.364335 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-qskdh" podStartSLOduration=71.532639261 podStartE2EDuration="1m24.364312212s" podCreationTimestamp="2025-11-24 17:17:57 +0000 UTC" firstStartedPulling="2025-11-24 17:19:06.64145234 +0000 UTC m=+1104.800647399" lastFinishedPulling="2025-11-24 17:19:19.473125301 +0000 UTC m=+1117.632320350" observedRunningTime="2025-11-24 17:19:21.330566013 +0000 UTC m=+1119.489761062" watchObservedRunningTime="2025-11-24 17:19:21.364312212 +0000 UTC m=+1119.523507261" Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.391224 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" podStartSLOduration=71.511001299 podStartE2EDuration="1m24.391204203s" podCreationTimestamp="2025-11-24 17:17:57 +0000 UTC" firstStartedPulling="2025-11-24 17:19:06.640880993 +0000 UTC m=+1104.800076052" lastFinishedPulling="2025-11-24 17:19:19.521083917 +0000 UTC m=+1117.680278956" observedRunningTime="2025-11-24 17:19:21.348663533 +0000 UTC m=+1119.507858582" watchObservedRunningTime="2025-11-24 17:19:21.391204203 +0000 UTC m=+1119.550399252" Nov 24 17:19:21 crc kubenswrapper[4777]: I1124 17:19:21.520923 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:21 crc kubenswrapper[4777]: E1124 17:19:21.521184 4777 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 17:19:21 crc kubenswrapper[4777]: E1124 17:19:21.521218 4777 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 17:19:21 crc kubenswrapper[4777]: E1124 17:19:21.521280 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift podName:9c412d58-6198-4254-932d-5eb90420f8ba nodeName:}" failed. No retries permitted until 2025-11-24 17:19:25.521260086 +0000 UTC m=+1123.680455135 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift") pod "swift-storage-0" (UID: "9c412d58-6198-4254-932d-5eb90420f8ba") : configmap "swift-ring-files" not found Nov 24 17:19:22 crc kubenswrapper[4777]: I1124 17:19:22.318505 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" event={"ID":"1aec5e66-69c2-404c-865a-0995f09fd435","Type":"ContainerStarted","Data":"207bc400026eaf3c17caef36ffb3f42501b6536cbd33ce382cb5ff363fe0c1cc"} Nov 24 17:19:22 crc kubenswrapper[4777]: I1124 17:19:22.318834 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:19:22 crc kubenswrapper[4777]: I1124 17:19:22.320373 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1d71a086-37e0-4409-9db9-66e6f1374406","Type":"ContainerStarted","Data":"b629eaee7d4f3b619f073f8c6899402058219301389b9f8ae40709bd83c2ac3b"} Nov 24 17:19:22 crc kubenswrapper[4777]: I1124 17:19:22.332556 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" Nov 24 17:19:22 crc kubenswrapper[4777]: I1124 17:19:22.333078 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-compactor-0" event={"ID":"e99dec76-e21f-4aab-bdc9-ae098391ee6e","Type":"ContainerStarted","Data":"08148bcf2260ad0489778dd995fa3e0193598d971da889bafd74357c7d6692a0"} Nov 24 17:19:22 crc kubenswrapper[4777]: I1124 17:19:22.404980 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-compactor-0" podStartSLOduration=71.991906999 podStartE2EDuration="1m25.404940293s" podCreationTimestamp="2025-11-24 17:17:57 +0000 UTC" firstStartedPulling="2025-11-24 17:19:06.559594481 +0000 UTC m=+1104.718789530" lastFinishedPulling="2025-11-24 17:19:19.972627775 +0000 UTC m=+1118.131822824" observedRunningTime="2025-11-24 17:19:22.391451396 +0000 UTC m=+1120.550646445" watchObservedRunningTime="2025-11-24 17:19:22.404940293 +0000 UTC m=+1120.564135342" Nov 24 17:19:22 crc kubenswrapper[4777]: I1124 17:19:22.408087 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-znzk8" podStartSLOduration=71.584266751 podStartE2EDuration="1m25.408075573s" podCreationTimestamp="2025-11-24 17:17:57 +0000 UTC" firstStartedPulling="2025-11-24 17:19:06.641979465 +0000 UTC m=+1104.801174534" lastFinishedPulling="2025-11-24 17:19:20.465788307 +0000 UTC m=+1118.624983356" observedRunningTime="2025-11-24 17:19:22.362154157 +0000 UTC m=+1120.521349216" watchObservedRunningTime="2025-11-24 17:19:22.408075573 +0000 UTC m=+1120.567270632" Nov 24 17:19:23 crc kubenswrapper[4777]: I1124 17:19:23.344295 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" event={"ID":"9c73a207-d21c-4157-b27a-41656a5f6af1","Type":"ContainerStarted","Data":"da8d65b37d9ddc2effc9d3c002a7bc8b3148bc39e67c950fa82dae9a4c92b5a9"} Nov 24 17:19:23 crc kubenswrapper[4777]: I1124 17:19:23.345002 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:19:23 crc kubenswrapper[4777]: I1124 17:19:23.347585 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f9ab961d-98f3-4a5e-9583-a168fa95f7b4","Type":"ContainerStarted","Data":"efb7c66a1887ce7a1d81fe309029105ee9be5adcb1357020783c58d1c95679c7"} Nov 24 17:19:23 crc kubenswrapper[4777]: I1124 17:19:23.350316 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-ingester-0" event={"ID":"f68bfbdc-e400-488d-9f32-ea90fd1160e3","Type":"ContainerStarted","Data":"341371233d1ca15e41a4f837f3aca6677af21e1a940cd591b04608e4473fadce"} Nov 24 17:19:23 crc kubenswrapper[4777]: I1124 17:19:23.350573 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:19:23 crc kubenswrapper[4777]: I1124 17:19:23.353980 4777 generic.go:334] "Generic (PLEG): container finished" podID="3370ce09-05a5-4d40-b89b-d75720b39677" containerID="6b091a1321d3320dd38a52f6501db9464a59367e9933462a9fdb1c64f7300ce2" exitCode=0 Nov 24 17:19:23 crc kubenswrapper[4777]: I1124 17:19:23.354148 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" event={"ID":"3370ce09-05a5-4d40-b89b-d75720b39677","Type":"ContainerDied","Data":"6b091a1321d3320dd38a52f6501db9464a59367e9933462a9fdb1c64f7300ce2"} Nov 24 17:19:23 crc kubenswrapper[4777]: I1124 17:19:23.354739 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:19:23 crc kubenswrapper[4777]: I1124 17:19:23.368658 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" podStartSLOduration=72.697978281 podStartE2EDuration="1m26.368625078s" podCreationTimestamp="2025-11-24 17:17:57 +0000 UTC" firstStartedPulling="2025-11-24 17:19:06.648471321 +0000 UTC m=+1104.807666380" lastFinishedPulling="2025-11-24 17:19:20.319118128 +0000 UTC m=+1118.478313177" observedRunningTime="2025-11-24 17:19:23.359589528 +0000 UTC m=+1121.518784597" watchObservedRunningTime="2025-11-24 17:19:23.368625078 +0000 UTC m=+1121.527820117" Nov 24 17:19:23 crc kubenswrapper[4777]: I1124 17:19:23.384765 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-ingester-0" podStartSLOduration=72.385778422 podStartE2EDuration="1m26.38473471s" podCreationTimestamp="2025-11-24 17:17:57 +0000 UTC" firstStartedPulling="2025-11-24 17:19:06.613263711 +0000 UTC m=+1104.772458760" lastFinishedPulling="2025-11-24 17:19:20.612219989 +0000 UTC m=+1118.771415048" observedRunningTime="2025-11-24 17:19:23.384659528 +0000 UTC m=+1121.543854587" watchObservedRunningTime="2025-11-24 17:19:23.38473471 +0000 UTC m=+1121.543929759" Nov 24 17:19:25 crc kubenswrapper[4777]: I1124 17:19:25.609600 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:25 crc kubenswrapper[4777]: E1124 17:19:25.609843 4777 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 17:19:25 crc kubenswrapper[4777]: E1124 17:19:25.609887 4777 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 17:19:25 crc kubenswrapper[4777]: E1124 17:19:25.610019 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift podName:9c412d58-6198-4254-932d-5eb90420f8ba nodeName:}" failed. No retries permitted until 2025-11-24 17:19:33.609948966 +0000 UTC m=+1131.769144055 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift") pod "swift-storage-0" (UID: "9c412d58-6198-4254-932d-5eb90420f8ba") : configmap "swift-ring-files" not found Nov 24 17:19:27 crc kubenswrapper[4777]: I1124 17:19:27.395620 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dkbds" event={"ID":"3641f728-c179-4f10-93c2-98f74763e77f","Type":"ContainerStarted","Data":"e5ac2d45aac848a6762adae453469fdebb06cabfd768888ef38f0bb5aa48f932"} Nov 24 17:19:28 crc kubenswrapper[4777]: I1124 17:19:28.404659 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4eba3f22-b7ad-464d-a0c9-437c449d124b","Type":"ContainerStarted","Data":"c7fca1f0840634419c786882684b6392087d49e7a864dce40ab2820ca03a60d0"} Nov 24 17:19:28 crc kubenswrapper[4777]: I1124 17:19:28.407358 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" event={"ID":"3370ce09-05a5-4d40-b89b-d75720b39677","Type":"ContainerStarted","Data":"313143c9d06733f32784b34ade782d8300249aec474b272a14114e2e6ec7ff71"} Nov 24 17:19:28 crc kubenswrapper[4777]: I1124 17:19:28.407537 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" Nov 24 17:19:28 crc kubenswrapper[4777]: I1124 17:19:28.409665 4777 generic.go:334] "Generic (PLEG): container finished" podID="3641f728-c179-4f10-93c2-98f74763e77f" containerID="e5ac2d45aac848a6762adae453469fdebb06cabfd768888ef38f0bb5aa48f932" exitCode=0 Nov 24 17:19:28 crc kubenswrapper[4777]: I1124 17:19:28.409710 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dkbds" event={"ID":"3641f728-c179-4f10-93c2-98f74763e77f","Type":"ContainerDied","Data":"e5ac2d45aac848a6762adae453469fdebb06cabfd768888ef38f0bb5aa48f932"} Nov 24 17:19:28 crc kubenswrapper[4777]: I1124 17:19:28.412931 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" event={"ID":"c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb","Type":"ContainerStarted","Data":"04ac8d364ae8eb5e46b44e74adcfdc4b10b08d4ad428f131df2b18961210c92d"} Nov 24 17:19:28 crc kubenswrapper[4777]: I1124 17:19:28.414369 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"155b4504-c156-4af8-a9e0-1dc66c9d7d43","Type":"ContainerStarted","Data":"f46b0fb98030f89ed777d3807dcbac2a515d8a96488e3f477fe4aa40b840f5c6"} Nov 24 17:19:28 crc kubenswrapper[4777]: I1124 17:19:28.415550 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-q4l2k" event={"ID":"73928e73-667c-4b69-aed9-72f32012fbdc","Type":"ContainerStarted","Data":"04aa53aa46222f67a32a1b8c6ac0ea35cc0b1d2a3ee7ce1669e6eaf0a9806662"} Nov 24 17:19:28 crc kubenswrapper[4777]: I1124 17:19:28.416910 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4ccca46a-1cac-4693-8490-2612b138c9d5","Type":"ContainerStarted","Data":"78512587e30a36de29f993b7d76e908ace85cfed73d7e045686b3929dee8f0ab"} Nov 24 17:19:28 crc kubenswrapper[4777]: I1124 17:19:28.418100 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" event={"ID":"54920d72-c01f-4d69-bc52-873f086d51c5","Type":"ContainerStarted","Data":"e07bc93936a31be4eedee42acf1852d267cc90d95a4d553a35711103c3e38625"} Nov 24 17:19:28 crc kubenswrapper[4777]: I1124 17:19:28.419683 4777 generic.go:334] "Generic (PLEG): container finished" podID="1d71a086-37e0-4409-9db9-66e6f1374406" containerID="b629eaee7d4f3b619f073f8c6899402058219301389b9f8ae40709bd83c2ac3b" exitCode=0 Nov 24 17:19:28 crc kubenswrapper[4777]: I1124 17:19:28.419710 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1d71a086-37e0-4409-9db9-66e6f1374406","Type":"ContainerDied","Data":"b629eaee7d4f3b619f073f8c6899402058219301389b9f8ae40709bd83c2ac3b"} Nov 24 17:19:28 crc kubenswrapper[4777]: I1124 17:19:28.454637 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" podStartSLOduration=10.195291884 podStartE2EDuration="1m48.454618457s" podCreationTimestamp="2025-11-24 17:17:40 +0000 UTC" firstStartedPulling="2025-11-24 17:17:41.241795511 +0000 UTC m=+1019.400990600" lastFinishedPulling="2025-11-24 17:19:19.501122084 +0000 UTC m=+1117.660317173" observedRunningTime="2025-11-24 17:19:28.429465756 +0000 UTC m=+1126.588660805" watchObservedRunningTime="2025-11-24 17:19:28.454618457 +0000 UTC m=+1126.613813506" Nov 24 17:19:29 crc kubenswrapper[4777]: I1124 17:19:29.432496 4777 generic.go:334] "Generic (PLEG): container finished" podID="54920d72-c01f-4d69-bc52-873f086d51c5" containerID="e07bc93936a31be4eedee42acf1852d267cc90d95a4d553a35711103c3e38625" exitCode=0 Nov 24 17:19:29 crc kubenswrapper[4777]: I1124 17:19:29.432740 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" event={"ID":"54920d72-c01f-4d69-bc52-873f086d51c5","Type":"ContainerDied","Data":"e07bc93936a31be4eedee42acf1852d267cc90d95a4d553a35711103c3e38625"} Nov 24 17:19:29 crc kubenswrapper[4777]: I1124 17:19:29.434658 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 17:19:29 crc kubenswrapper[4777]: I1124 17:19:29.434957 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-q4l2k" Nov 24 17:19:29 crc kubenswrapper[4777]: I1124 17:19:29.435249 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:19:29 crc kubenswrapper[4777]: I1124 17:19:29.504736 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-q4l2k" podStartSLOduration=84.940833762 podStartE2EDuration="1m39.504716272s" podCreationTimestamp="2025-11-24 17:17:50 +0000 UTC" firstStartedPulling="2025-11-24 17:19:05.902736981 +0000 UTC m=+1104.061932030" lastFinishedPulling="2025-11-24 17:19:20.466619481 +0000 UTC m=+1118.625814540" observedRunningTime="2025-11-24 17:19:29.497100913 +0000 UTC m=+1127.656295972" watchObservedRunningTime="2025-11-24 17:19:29.504716272 +0000 UTC m=+1127.663911331" Nov 24 17:19:29 crc kubenswrapper[4777]: I1124 17:19:29.524422 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=82.079458984 podStartE2EDuration="1m43.524395335s" podCreationTimestamp="2025-11-24 17:17:46 +0000 UTC" firstStartedPulling="2025-11-24 17:19:06.064558395 +0000 UTC m=+1104.223753444" lastFinishedPulling="2025-11-24 17:19:27.509494746 +0000 UTC m=+1125.668689795" observedRunningTime="2025-11-24 17:19:29.513870244 +0000 UTC m=+1127.673065303" watchObservedRunningTime="2025-11-24 17:19:29.524395335 +0000 UTC m=+1127.683590394" Nov 24 17:19:29 crc kubenswrapper[4777]: I1124 17:19:29.567715 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" podStartSLOduration=78.675005449 podStartE2EDuration="1m32.567690908s" podCreationTimestamp="2025-11-24 17:17:57 +0000 UTC" firstStartedPulling="2025-11-24 17:19:06.800515284 +0000 UTC m=+1104.959710333" lastFinishedPulling="2025-11-24 17:19:20.693200743 +0000 UTC m=+1118.852395792" observedRunningTime="2025-11-24 17:19:29.562035776 +0000 UTC m=+1127.721230865" watchObservedRunningTime="2025-11-24 17:19:29.567690908 +0000 UTC m=+1127.726885967" Nov 24 17:19:30 crc kubenswrapper[4777]: I1124 17:19:30.443090 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4c707277-46e4-42ea-9076-566b2d01f544","Type":"ContainerStarted","Data":"9dfcd853b3fc5572da0af858e5836f1849e0716f5fca857126f422e455bcc6b7"} Nov 24 17:19:30 crc kubenswrapper[4777]: I1124 17:19:30.448158 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7163366e-9a39-4379-af0a-d6c7c998ab7f","Type":"ContainerStarted","Data":"51d916dedea9f0bddc55f0a8a97bf14ba1e65ae0a91b751dee93cb87d18481b8"} Nov 24 17:19:33 crc kubenswrapper[4777]: I1124 17:19:33.673716 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:33 crc kubenswrapper[4777]: E1124 17:19:33.674050 4777 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 17:19:33 crc kubenswrapper[4777]: E1124 17:19:33.674342 4777 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 17:19:33 crc kubenswrapper[4777]: E1124 17:19:33.674441 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift podName:9c412d58-6198-4254-932d-5eb90420f8ba nodeName:}" failed. No retries permitted until 2025-11-24 17:19:49.674408006 +0000 UTC m=+1147.833603085 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift") pod "swift-storage-0" (UID: "9c412d58-6198-4254-932d-5eb90420f8ba") : configmap "swift-ring-files" not found Nov 24 17:19:34 crc kubenswrapper[4777]: I1124 17:19:34.528938 4777 generic.go:334] "Generic (PLEG): container finished" podID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerID="efb7c66a1887ce7a1d81fe309029105ee9be5adcb1357020783c58d1c95679c7" exitCode=0 Nov 24 17:19:34 crc kubenswrapper[4777]: I1124 17:19:34.529266 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f9ab961d-98f3-4a5e-9583-a168fa95f7b4","Type":"ContainerDied","Data":"efb7c66a1887ce7a1d81fe309029105ee9be5adcb1357020783c58d1c95679c7"} Nov 24 17:19:35 crc kubenswrapper[4777]: I1124 17:19:35.741108 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" Nov 24 17:19:36 crc kubenswrapper[4777]: I1124 17:19:36.451049 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 17:19:37 crc kubenswrapper[4777]: I1124 17:19:37.856383 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-x7zhc" Nov 24 17:19:38 crc kubenswrapper[4777]: I1124 17:19:38.725108 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="f68bfbdc-e400-488d-9f32-ea90fd1160e3" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 17:19:38 crc kubenswrapper[4777]: I1124 17:19:38.828880 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-compactor-0" Nov 24 17:19:39 crc kubenswrapper[4777]: I1124 17:19:39.042158 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-index-gateway-0" Nov 24 17:19:40 crc kubenswrapper[4777]: I1124 17:19:40.594456 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"055503f1-ab1c-47db-be36-37a3f572be32","Type":"ContainerStarted","Data":"3b8539103958963891076ac10e2c1bad4ff367734abaad602c12e6f15ab102e2"} Nov 24 17:19:44 crc kubenswrapper[4777]: E1124 17:19:44.162445 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified" Nov 24 17:19:44 crc kubenswrapper[4777]: E1124 17:19:44.163493 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:swift-ring-rebalance,Image:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,Command:[/usr/local/bin/swift-ring-tool all],Args:[],WorkingDir:/etc/swift,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CM_NAME,Value:swift-ring-files,ValueFrom:nil,},EnvVar{Name:NAMESPACE,Value:openstack,ValueFrom:nil,},EnvVar{Name:OWNER_APIVERSION,Value:swift.openstack.org/v1beta1,ValueFrom:nil,},EnvVar{Name:OWNER_KIND,Value:SwiftRing,ValueFrom:nil,},EnvVar{Name:OWNER_NAME,Value:swift-ring,ValueFrom:nil,},EnvVar{Name:OWNER_UID,Value:ce6dd77a-b175-4e9c-9c21-3a11825cc1d0,ValueFrom:nil,},EnvVar{Name:SWIFT_MIN_PART_HOURS,Value:1,ValueFrom:nil,},EnvVar{Name:SWIFT_PART_POWER,Value:10,ValueFrom:nil,},EnvVar{Name:SWIFT_REPLICAS,Value:1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/swift-ring-tool,SubPath:swift-ring-tool,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:swiftconf,ReadOnly:true,MountPath:/etc/swift/swift.conf,SubPath:swift.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-swift,ReadOnly:false,MountPath:/etc/swift,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ring-data-devices,ReadOnly:true,MountPath:/var/lib/config-data/ring-devices,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dispersionconf,ReadOnly:true,MountPath:/etc/swift/dispersion.conf,SubPath:dispersion.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-75kwr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42445,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-ring-rebalance-xgm6s_openstack(3ec24d0c-556b-4a24-8f42-38fed08efe6b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:19:44 crc kubenswrapper[4777]: E1124 17:19:44.166423 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"swift-ring-rebalance\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/swift-ring-rebalance-xgm6s" podUID="3ec24d0c-556b-4a24-8f42-38fed08efe6b" Nov 24 17:19:44 crc kubenswrapper[4777]: E1124 17:19:44.239914 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="737d5505-10bc-4e56-b846-01a3af071b38" Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.270605 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.270668 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.270705 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.271400 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3cf02e0241e501ffd35b75c9b5e93243f3fad1bfd50b46e463488c0fec7f7c53"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.271467 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://3cf02e0241e501ffd35b75c9b5e93243f3fad1bfd50b46e463488c0fec7f7c53" gracePeriod=600 Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.638351 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dkbds" event={"ID":"3641f728-c179-4f10-93c2-98f74763e77f","Type":"ContainerStarted","Data":"b285cbcebba67f60a00ed75500bd530a8f84cb485b550ef39764d89debdbeeed"} Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.640358 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"737d5505-10bc-4e56-b846-01a3af071b38","Type":"ContainerStarted","Data":"4171069229d45c4f80263ebd8e7cd96be024a697cf5cc9d677616d97483337fb"} Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.642705 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="3cf02e0241e501ffd35b75c9b5e93243f3fad1bfd50b46e463488c0fec7f7c53" exitCode=0 Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.642749 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"3cf02e0241e501ffd35b75c9b5e93243f3fad1bfd50b46e463488c0fec7f7c53"} Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.642822 4777 scope.go:117] "RemoveContainer" containerID="90d22c4d0e8b2ba20584ba22b6a107d623c9ded1ab193fb10e43b251d31391cc" Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.645372 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4ccca46a-1cac-4693-8490-2612b138c9d5","Type":"ContainerStarted","Data":"8629edbce08119c7d671c800ec8ad66f273be18757fe7a2c72af4fae080d2db5"} Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.649361 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" event={"ID":"54920d72-c01f-4d69-bc52-873f086d51c5","Type":"ContainerStarted","Data":"c9db071a672e8c62c9c36e2161fb11ce701bafce5a37b2b8566f4b02ee4ccafd"} Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.649430 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.651861 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1d71a086-37e0-4409-9db9-66e6f1374406","Type":"ContainerStarted","Data":"ffc808c047b92c6ad977968076d5a7f0e45a03ee738f52246b7cebccb4552aa9"} Nov 24 17:19:44 crc kubenswrapper[4777]: E1124 17:19:44.655206 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"swift-ring-rebalance\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified\\\"\"" pod="openstack/swift-ring-rebalance-xgm6s" podUID="3ec24d0c-556b-4a24-8f42-38fed08efe6b" Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.683680 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.702488 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" podStartSLOduration=28.702464771 podStartE2EDuration="28.702464771s" podCreationTimestamp="2025-11-24 17:19:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:19:44.699161926 +0000 UTC m=+1142.858356995" watchObservedRunningTime="2025-11-24 17:19:44.702464771 +0000 UTC m=+1142.861659820" Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.730105 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.775441 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=98.742509662 podStartE2EDuration="1m52.775418954s" podCreationTimestamp="2025-11-24 17:17:52 +0000 UTC" firstStartedPulling="2025-11-24 17:19:06.686262116 +0000 UTC m=+1104.845457155" lastFinishedPulling="2025-11-24 17:19:20.719171398 +0000 UTC m=+1118.878366447" observedRunningTime="2025-11-24 17:19:44.747290757 +0000 UTC m=+1142.906485846" watchObservedRunningTime="2025-11-24 17:19:44.775418954 +0000 UTC m=+1142.934614003" Nov 24 17:19:44 crc kubenswrapper[4777]: I1124 17:19:44.797056 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=27.753893739 podStartE2EDuration="2m3.797028964s" podCreationTimestamp="2025-11-24 17:17:41 +0000 UTC" firstStartedPulling="2025-11-24 17:17:43.971152346 +0000 UTC m=+1022.130347395" lastFinishedPulling="2025-11-24 17:19:20.014287571 +0000 UTC m=+1118.173482620" observedRunningTime="2025-11-24 17:19:44.777333029 +0000 UTC m=+1142.936528068" watchObservedRunningTime="2025-11-24 17:19:44.797028964 +0000 UTC m=+1142.956224013" Nov 24 17:19:45 crc kubenswrapper[4777]: I1124 17:19:45.665094 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dkbds" event={"ID":"3641f728-c179-4f10-93c2-98f74763e77f","Type":"ContainerStarted","Data":"12520217990ef895222fcd77237979e39a9601c499916a3e43a17c2375f89a04"} Nov 24 17:19:45 crc kubenswrapper[4777]: I1124 17:19:45.665494 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:19:45 crc kubenswrapper[4777]: I1124 17:19:45.666360 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:19:45 crc kubenswrapper[4777]: I1124 17:19:45.667805 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"a6faf411f1a27fb01d7bc7ffb8b05ff8c95006e6030cf4cec387d5e3b99354f9"} Nov 24 17:19:45 crc kubenswrapper[4777]: I1124 17:19:45.668710 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 24 17:19:45 crc kubenswrapper[4777]: I1124 17:19:45.697600 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-dkbds" podStartSLOduration=101.862970925 podStartE2EDuration="1m55.697576817s" podCreationTimestamp="2025-11-24 17:17:50 +0000 UTC" firstStartedPulling="2025-11-24 17:19:06.174892271 +0000 UTC m=+1104.334087320" lastFinishedPulling="2025-11-24 17:19:20.009498163 +0000 UTC m=+1118.168693212" observedRunningTime="2025-11-24 17:19:45.683480352 +0000 UTC m=+1143.842675401" watchObservedRunningTime="2025-11-24 17:19:45.697576817 +0000 UTC m=+1143.856771866" Nov 24 17:19:45 crc kubenswrapper[4777]: I1124 17:19:45.727836 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 24 17:19:45 crc kubenswrapper[4777]: I1124 17:19:45.998862 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-n7hxb"] Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.042821 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-kzbfv"] Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.044411 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.047364 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.051981 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-lgxzj"] Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.053602 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.060182 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.079681 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-kzbfv"] Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.091946 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lgxzj"] Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.151849 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/04bed066-31b6-49d0-90ee-68e38f7944c1-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.151933 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-dns-svc\") pod \"dnsmasq-dns-57d65f699f-kzbfv\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.151984 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-config\") pod \"dnsmasq-dns-57d65f699f-kzbfv\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.152004 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04bed066-31b6-49d0-90ee-68e38f7944c1-config\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.152018 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04bed066-31b6-49d0-90ee-68e38f7944c1-combined-ca-bundle\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.152036 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mknf2\" (UniqueName: \"kubernetes.io/projected/db510872-2747-447b-9177-e50ae5043933-kube-api-access-mknf2\") pod \"dnsmasq-dns-57d65f699f-kzbfv\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.152053 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/04bed066-31b6-49d0-90ee-68e38f7944c1-ovn-rundir\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.152102 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjdxx\" (UniqueName: \"kubernetes.io/projected/04bed066-31b6-49d0-90ee-68e38f7944c1-kube-api-access-jjdxx\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.152121 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/04bed066-31b6-49d0-90ee-68e38f7944c1-ovs-rundir\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.152210 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-kzbfv\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.210378 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-kzbfv"] Nov 24 17:19:46 crc kubenswrapper[4777]: E1124 17:19:46.211002 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-mknf2 ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" podUID="db510872-2747-447b-9177-e50ae5043933" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.237330 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-fq2lk"] Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.238941 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.241060 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.253186 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-dns-svc\") pod \"dnsmasq-dns-57d65f699f-kzbfv\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.253243 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-config\") pod \"dnsmasq-dns-57d65f699f-kzbfv\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.253269 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04bed066-31b6-49d0-90ee-68e38f7944c1-config\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.253284 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04bed066-31b6-49d0-90ee-68e38f7944c1-combined-ca-bundle\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.253304 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mknf2\" (UniqueName: \"kubernetes.io/projected/db510872-2747-447b-9177-e50ae5043933-kube-api-access-mknf2\") pod \"dnsmasq-dns-57d65f699f-kzbfv\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.253325 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/04bed066-31b6-49d0-90ee-68e38f7944c1-ovn-rundir\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.253353 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjdxx\" (UniqueName: \"kubernetes.io/projected/04bed066-31b6-49d0-90ee-68e38f7944c1-kube-api-access-jjdxx\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.253368 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/04bed066-31b6-49d0-90ee-68e38f7944c1-ovs-rundir\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.253439 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-kzbfv\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.253481 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/04bed066-31b6-49d0-90ee-68e38f7944c1-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.253663 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/04bed066-31b6-49d0-90ee-68e38f7944c1-ovn-rundir\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.254007 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/04bed066-31b6-49d0-90ee-68e38f7944c1-ovs-rundir\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.254121 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-dns-svc\") pod \"dnsmasq-dns-57d65f699f-kzbfv\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.254225 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-config\") pod \"dnsmasq-dns-57d65f699f-kzbfv\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.254240 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04bed066-31b6-49d0-90ee-68e38f7944c1-config\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.261765 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-kzbfv\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.262364 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/04bed066-31b6-49d0-90ee-68e38f7944c1-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.278949 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04bed066-31b6-49d0-90ee-68e38f7944c1-combined-ca-bundle\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.293669 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjdxx\" (UniqueName: \"kubernetes.io/projected/04bed066-31b6-49d0-90ee-68e38f7944c1-kube-api-access-jjdxx\") pod \"ovn-controller-metrics-lgxzj\" (UID: \"04bed066-31b6-49d0-90ee-68e38f7944c1\") " pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.294599 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mknf2\" (UniqueName: \"kubernetes.io/projected/db510872-2747-447b-9177-e50ae5043933-kube-api-access-mknf2\") pod \"dnsmasq-dns-57d65f699f-kzbfv\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.306706 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-fq2lk"] Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.355578 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-fq2lk\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.355628 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-fq2lk\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.355646 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-config\") pod \"dnsmasq-dns-b8fbc5445-fq2lk\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.355749 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-fq2lk\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.355767 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42r8s\" (UniqueName: \"kubernetes.io/projected/10e17cd8-dd7d-476a-96f7-27ac1f938b83-kube-api-access-42r8s\") pod \"dnsmasq-dns-b8fbc5445-fq2lk\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.387792 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lgxzj" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.456767 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-fq2lk\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.456821 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-fq2lk\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.456846 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-config\") pod \"dnsmasq-dns-b8fbc5445-fq2lk\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.456986 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-fq2lk\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.457009 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42r8s\" (UniqueName: \"kubernetes.io/projected/10e17cd8-dd7d-476a-96f7-27ac1f938b83-kube-api-access-42r8s\") pod \"dnsmasq-dns-b8fbc5445-fq2lk\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.458287 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-fq2lk\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.458523 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-config\") pod \"dnsmasq-dns-b8fbc5445-fq2lk\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.458540 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-fq2lk\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.459229 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-fq2lk\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.479838 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42r8s\" (UniqueName: \"kubernetes.io/projected/10e17cd8-dd7d-476a-96f7-27ac1f938b83-kube-api-access-42r8s\") pod \"dnsmasq-dns-b8fbc5445-fq2lk\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.554435 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.689654 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" podUID="54920d72-c01f-4d69-bc52-873f086d51c5" containerName="dnsmasq-dns" containerID="cri-o://c9db071a672e8c62c9c36e2161fb11ce701bafce5a37b2b8566f4b02ee4ccafd" gracePeriod=10 Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.690500 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.708657 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.764471 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-ovsdbserver-nb\") pod \"db510872-2747-447b-9177-e50ae5043933\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.764556 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mknf2\" (UniqueName: \"kubernetes.io/projected/db510872-2747-447b-9177-e50ae5043933-kube-api-access-mknf2\") pod \"db510872-2747-447b-9177-e50ae5043933\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.764619 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-dns-svc\") pod \"db510872-2747-447b-9177-e50ae5043933\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.764649 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-config\") pod \"db510872-2747-447b-9177-e50ae5043933\" (UID: \"db510872-2747-447b-9177-e50ae5043933\") " Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.764938 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "db510872-2747-447b-9177-e50ae5043933" (UID: "db510872-2747-447b-9177-e50ae5043933"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.765163 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.765940 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-config" (OuterVolumeSpecName: "config") pod "db510872-2747-447b-9177-e50ae5043933" (UID: "db510872-2747-447b-9177-e50ae5043933"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.767084 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "db510872-2747-447b-9177-e50ae5043933" (UID: "db510872-2747-447b-9177-e50ae5043933"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.770267 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db510872-2747-447b-9177-e50ae5043933-kube-api-access-mknf2" (OuterVolumeSpecName: "kube-api-access-mknf2") pod "db510872-2747-447b-9177-e50ae5043933" (UID: "db510872-2747-447b-9177-e50ae5043933"). InnerVolumeSpecName "kube-api-access-mknf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.827833 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lgxzj"] Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.867001 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.867032 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db510872-2747-447b-9177-e50ae5043933-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:46 crc kubenswrapper[4777]: I1124 17:19:46.867045 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mknf2\" (UniqueName: \"kubernetes.io/projected/db510872-2747-447b-9177-e50ae5043933-kube-api-access-mknf2\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.600365 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jg9l8" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.637489 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.685953 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54920d72-c01f-4d69-bc52-873f086d51c5-dns-svc\") pod \"54920d72-c01f-4d69-bc52-873f086d51c5\" (UID: \"54920d72-c01f-4d69-bc52-873f086d51c5\") " Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.686035 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n45n\" (UniqueName: \"kubernetes.io/projected/54920d72-c01f-4d69-bc52-873f086d51c5-kube-api-access-2n45n\") pod \"54920d72-c01f-4d69-bc52-873f086d51c5\" (UID: \"54920d72-c01f-4d69-bc52-873f086d51c5\") " Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.686056 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54920d72-c01f-4d69-bc52-873f086d51c5-config\") pod \"54920d72-c01f-4d69-bc52-873f086d51c5\" (UID: \"54920d72-c01f-4d69-bc52-873f086d51c5\") " Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.690855 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54920d72-c01f-4d69-bc52-873f086d51c5-kube-api-access-2n45n" (OuterVolumeSpecName: "kube-api-access-2n45n") pod "54920d72-c01f-4d69-bc52-873f086d51c5" (UID: "54920d72-c01f-4d69-bc52-873f086d51c5"). InnerVolumeSpecName "kube-api-access-2n45n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.702632 4777 generic.go:334] "Generic (PLEG): container finished" podID="055503f1-ab1c-47db-be36-37a3f572be32" containerID="3b8539103958963891076ac10e2c1bad4ff367734abaad602c12e6f15ab102e2" exitCode=0 Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.702682 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"055503f1-ab1c-47db-be36-37a3f572be32","Type":"ContainerDied","Data":"3b8539103958963891076ac10e2c1bad4ff367734abaad602c12e6f15ab102e2"} Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.729727 4777 generic.go:334] "Generic (PLEG): container finished" podID="54920d72-c01f-4d69-bc52-873f086d51c5" containerID="c9db071a672e8c62c9c36e2161fb11ce701bafce5a37b2b8566f4b02ee4ccafd" exitCode=0 Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.729796 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" event={"ID":"54920d72-c01f-4d69-bc52-873f086d51c5","Type":"ContainerDied","Data":"c9db071a672e8c62c9c36e2161fb11ce701bafce5a37b2b8566f4b02ee4ccafd"} Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.729828 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" event={"ID":"54920d72-c01f-4d69-bc52-873f086d51c5","Type":"ContainerDied","Data":"8c6184ef938e3c4c8cd1a856d68d0c35cacd0bde0e24829cfb1d971647bdc727"} Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.729847 4777 scope.go:117] "RemoveContainer" containerID="c9db071a672e8c62c9c36e2161fb11ce701bafce5a37b2b8566f4b02ee4ccafd" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.730002 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-n7hxb" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.738148 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-kzbfv" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.738247 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lgxzj" event={"ID":"04bed066-31b6-49d0-90ee-68e38f7944c1","Type":"ContainerStarted","Data":"a049546ff7c10392efb57064d07e6aa3b82cfa98011bc40bcfa5953be3050755"} Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.738840 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54920d72-c01f-4d69-bc52-873f086d51c5-config" (OuterVolumeSpecName: "config") pod "54920d72-c01f-4d69-bc52-873f086d51c5" (UID: "54920d72-c01f-4d69-bc52-873f086d51c5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.752287 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54920d72-c01f-4d69-bc52-873f086d51c5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "54920d72-c01f-4d69-bc52-873f086d51c5" (UID: "54920d72-c01f-4d69-bc52-873f086d51c5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.780074 4777 scope.go:117] "RemoveContainer" containerID="e07bc93936a31be4eedee42acf1852d267cc90d95a4d553a35711103c3e38625" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.791575 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54920d72-c01f-4d69-bc52-873f086d51c5-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.791601 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n45n\" (UniqueName: \"kubernetes.io/projected/54920d72-c01f-4d69-bc52-873f086d51c5-kube-api-access-2n45n\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.791610 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54920d72-c01f-4d69-bc52-873f086d51c5-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.792923 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-kzbfv"] Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.799461 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-kzbfv"] Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.801657 4777 scope.go:117] "RemoveContainer" containerID="c9db071a672e8c62c9c36e2161fb11ce701bafce5a37b2b8566f4b02ee4ccafd" Nov 24 17:19:47 crc kubenswrapper[4777]: E1124 17:19:47.802319 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9db071a672e8c62c9c36e2161fb11ce701bafce5a37b2b8566f4b02ee4ccafd\": container with ID starting with c9db071a672e8c62c9c36e2161fb11ce701bafce5a37b2b8566f4b02ee4ccafd not found: ID does not exist" containerID="c9db071a672e8c62c9c36e2161fb11ce701bafce5a37b2b8566f4b02ee4ccafd" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.802360 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9db071a672e8c62c9c36e2161fb11ce701bafce5a37b2b8566f4b02ee4ccafd"} err="failed to get container status \"c9db071a672e8c62c9c36e2161fb11ce701bafce5a37b2b8566f4b02ee4ccafd\": rpc error: code = NotFound desc = could not find container \"c9db071a672e8c62c9c36e2161fb11ce701bafce5a37b2b8566f4b02ee4ccafd\": container with ID starting with c9db071a672e8c62c9c36e2161fb11ce701bafce5a37b2b8566f4b02ee4ccafd not found: ID does not exist" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.802389 4777 scope.go:117] "RemoveContainer" containerID="e07bc93936a31be4eedee42acf1852d267cc90d95a4d553a35711103c3e38625" Nov 24 17:19:47 crc kubenswrapper[4777]: E1124 17:19:47.803122 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e07bc93936a31be4eedee42acf1852d267cc90d95a4d553a35711103c3e38625\": container with ID starting with e07bc93936a31be4eedee42acf1852d267cc90d95a4d553a35711103c3e38625 not found: ID does not exist" containerID="e07bc93936a31be4eedee42acf1852d267cc90d95a4d553a35711103c3e38625" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.803157 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e07bc93936a31be4eedee42acf1852d267cc90d95a4d553a35711103c3e38625"} err="failed to get container status \"e07bc93936a31be4eedee42acf1852d267cc90d95a4d553a35711103c3e38625\": rpc error: code = NotFound desc = could not find container \"e07bc93936a31be4eedee42acf1852d267cc90d95a4d553a35711103c3e38625\": container with ID starting with e07bc93936a31be4eedee42acf1852d267cc90d95a4d553a35711103c3e38625 not found: ID does not exist" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.809333 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-querier-548665d79b-tqjct" Nov 24 17:19:47 crc kubenswrapper[4777]: I1124 17:19:47.860393 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-fq2lk"] Nov 24 17:19:48 crc kubenswrapper[4777]: I1124 17:19:48.065479 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-n7hxb"] Nov 24 17:19:48 crc kubenswrapper[4777]: I1124 17:19:48.071901 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-n7hxb"] Nov 24 17:19:48 crc kubenswrapper[4777]: I1124 17:19:48.720079 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="f68bfbdc-e400-488d-9f32-ea90fd1160e3" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 17:19:48 crc kubenswrapper[4777]: I1124 17:19:48.752855 4777 generic.go:334] "Generic (PLEG): container finished" podID="10e17cd8-dd7d-476a-96f7-27ac1f938b83" containerID="8fff73dbee5b4eb854cbde38f25bb95e59a9bec9ffcf68c8fe6681e1bc1e3467" exitCode=0 Nov 24 17:19:48 crc kubenswrapper[4777]: I1124 17:19:48.752905 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" event={"ID":"10e17cd8-dd7d-476a-96f7-27ac1f938b83","Type":"ContainerDied","Data":"8fff73dbee5b4eb854cbde38f25bb95e59a9bec9ffcf68c8fe6681e1bc1e3467"} Nov 24 17:19:48 crc kubenswrapper[4777]: I1124 17:19:48.753000 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" event={"ID":"10e17cd8-dd7d-476a-96f7-27ac1f938b83","Type":"ContainerStarted","Data":"4648cae05bb4802db0f8d420331a95ab758759c73ef7b24a03a2d81dfdce3e88"} Nov 24 17:19:48 crc kubenswrapper[4777]: I1124 17:19:48.756923 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lgxzj" event={"ID":"04bed066-31b6-49d0-90ee-68e38f7944c1","Type":"ContainerStarted","Data":"04389fe88dd19f085aa92a8d86031505fd99ecccfdc4e7546acf5444fda486fd"} Nov 24 17:19:48 crc kubenswrapper[4777]: I1124 17:19:48.759929 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"737d5505-10bc-4e56-b846-01a3af071b38","Type":"ContainerStarted","Data":"531724a676365ff23111e301925c420eebe9ab785cbe6379340c1c11131308d6"} Nov 24 17:19:48 crc kubenswrapper[4777]: I1124 17:19:48.796855 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-lgxzj" podStartSLOduration=2.796834384 podStartE2EDuration="2.796834384s" podCreationTimestamp="2025-11-24 17:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:19:48.78971645 +0000 UTC m=+1146.948911519" watchObservedRunningTime="2025-11-24 17:19:48.796834384 +0000 UTC m=+1146.956029433" Nov 24 17:19:48 crc kubenswrapper[4777]: I1124 17:19:48.818888 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=78.288861607 podStartE2EDuration="1m59.818863466s" podCreationTimestamp="2025-11-24 17:17:49 +0000 UTC" firstStartedPulling="2025-11-24 17:19:05.766476501 +0000 UTC m=+1103.925671550" lastFinishedPulling="2025-11-24 17:19:47.29647836 +0000 UTC m=+1145.455673409" observedRunningTime="2025-11-24 17:19:48.811449314 +0000 UTC m=+1146.970644373" watchObservedRunningTime="2025-11-24 17:19:48.818863466 +0000 UTC m=+1146.978058515" Nov 24 17:19:49 crc kubenswrapper[4777]: I1124 17:19:49.257032 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54920d72-c01f-4d69-bc52-873f086d51c5" path="/var/lib/kubelet/pods/54920d72-c01f-4d69-bc52-873f086d51c5/volumes" Nov 24 17:19:49 crc kubenswrapper[4777]: I1124 17:19:49.258083 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db510872-2747-447b-9177-e50ae5043933" path="/var/lib/kubelet/pods/db510872-2747-447b-9177-e50ae5043933/volumes" Nov 24 17:19:49 crc kubenswrapper[4777]: I1124 17:19:49.728449 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:19:49 crc kubenswrapper[4777]: E1124 17:19:49.728623 4777 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 17:19:49 crc kubenswrapper[4777]: E1124 17:19:49.728654 4777 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 17:19:49 crc kubenswrapper[4777]: E1124 17:19:49.728708 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift podName:9c412d58-6198-4254-932d-5eb90420f8ba nodeName:}" failed. No retries permitted until 2025-11-24 17:20:21.728689325 +0000 UTC m=+1179.887884374 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift") pod "swift-storage-0" (UID: "9c412d58-6198-4254-932d-5eb90420f8ba") : configmap "swift-ring-files" not found Nov 24 17:19:51 crc kubenswrapper[4777]: I1124 17:19:51.207075 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 24 17:19:51 crc kubenswrapper[4777]: I1124 17:19:51.207506 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 24 17:19:51 crc kubenswrapper[4777]: I1124 17:19:51.262176 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 24 17:19:51 crc kubenswrapper[4777]: I1124 17:19:51.789036 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f9ab961d-98f3-4a5e-9583-a168fa95f7b4","Type":"ContainerStarted","Data":"8bc55352e58db3657ba1188879f371842295e1d4d4f5233b6c23347da8f5c05e"} Nov 24 17:19:51 crc kubenswrapper[4777]: I1124 17:19:51.792058 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" event={"ID":"10e17cd8-dd7d-476a-96f7-27ac1f938b83","Type":"ContainerStarted","Data":"e0c118edd84e3687eb01c0c0fec0d5fce1f55f30cf6342ad71f957d16cd34ea7"} Nov 24 17:19:51 crc kubenswrapper[4777]: I1124 17:19:51.792211 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:51 crc kubenswrapper[4777]: I1124 17:19:51.794388 4777 generic.go:334] "Generic (PLEG): container finished" podID="4eba3f22-b7ad-464d-a0c9-437c449d124b" containerID="c7fca1f0840634419c786882684b6392087d49e7a864dce40ab2820ca03a60d0" exitCode=0 Nov 24 17:19:51 crc kubenswrapper[4777]: I1124 17:19:51.794687 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4eba3f22-b7ad-464d-a0c9-437c449d124b","Type":"ContainerDied","Data":"c7fca1f0840634419c786882684b6392087d49e7a864dce40ab2820ca03a60d0"} Nov 24 17:19:51 crc kubenswrapper[4777]: I1124 17:19:51.815084 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" podStartSLOduration=5.815067266 podStartE2EDuration="5.815067266s" podCreationTimestamp="2025-11-24 17:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:19:51.808418866 +0000 UTC m=+1149.967613915" watchObservedRunningTime="2025-11-24 17:19:51.815067266 +0000 UTC m=+1149.974262305" Nov 24 17:19:53 crc kubenswrapper[4777]: I1124 17:19:53.477238 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 24 17:19:53 crc kubenswrapper[4777]: I1124 17:19:53.479103 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 24 17:19:53 crc kubenswrapper[4777]: I1124 17:19:53.594683 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 24 17:19:53 crc kubenswrapper[4777]: I1124 17:19:53.941742 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.592085 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7311-account-create-xpsjd"] Nov 24 17:19:54 crc kubenswrapper[4777]: E1124 17:19:54.592715 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54920d72-c01f-4d69-bc52-873f086d51c5" containerName="dnsmasq-dns" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.592729 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="54920d72-c01f-4d69-bc52-873f086d51c5" containerName="dnsmasq-dns" Nov 24 17:19:54 crc kubenswrapper[4777]: E1124 17:19:54.592745 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54920d72-c01f-4d69-bc52-873f086d51c5" containerName="init" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.592752 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="54920d72-c01f-4d69-bc52-873f086d51c5" containerName="init" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.593006 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="54920d72-c01f-4d69-bc52-873f086d51c5" containerName="dnsmasq-dns" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.593617 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7311-account-create-xpsjd" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.596175 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.617668 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7311-account-create-xpsjd"] Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.669586 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-jdldd"] Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.671228 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jdldd" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.691577 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jdldd"] Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.754102 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a11e0baf-fea0-409a-9a16-1a467f83c1bf-operator-scripts\") pod \"keystone-7311-account-create-xpsjd\" (UID: \"a11e0baf-fea0-409a-9a16-1a467f83c1bf\") " pod="openstack/keystone-7311-account-create-xpsjd" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.754385 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm66c\" (UniqueName: \"kubernetes.io/projected/a11e0baf-fea0-409a-9a16-1a467f83c1bf-kube-api-access-dm66c\") pod \"keystone-7311-account-create-xpsjd\" (UID: \"a11e0baf-fea0-409a-9a16-1a467f83c1bf\") " pod="openstack/keystone-7311-account-create-xpsjd" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.828783 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-mxkvt"] Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.829897 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mxkvt" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.847409 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-mxkvt"] Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.856229 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f254c75d-4cc6-43b0-8f88-517f7539d2a5-operator-scripts\") pod \"keystone-db-create-jdldd\" (UID: \"f254c75d-4cc6-43b0-8f88-517f7539d2a5\") " pod="openstack/keystone-db-create-jdldd" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.856305 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a11e0baf-fea0-409a-9a16-1a467f83c1bf-operator-scripts\") pod \"keystone-7311-account-create-xpsjd\" (UID: \"a11e0baf-fea0-409a-9a16-1a467f83c1bf\") " pod="openstack/keystone-7311-account-create-xpsjd" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.856527 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm66c\" (UniqueName: \"kubernetes.io/projected/a11e0baf-fea0-409a-9a16-1a467f83c1bf-kube-api-access-dm66c\") pod \"keystone-7311-account-create-xpsjd\" (UID: \"a11e0baf-fea0-409a-9a16-1a467f83c1bf\") " pod="openstack/keystone-7311-account-create-xpsjd" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.856740 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgrh8\" (UniqueName: \"kubernetes.io/projected/f254c75d-4cc6-43b0-8f88-517f7539d2a5-kube-api-access-xgrh8\") pod \"keystone-db-create-jdldd\" (UID: \"f254c75d-4cc6-43b0-8f88-517f7539d2a5\") " pod="openstack/keystone-db-create-jdldd" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.857998 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a11e0baf-fea0-409a-9a16-1a467f83c1bf-operator-scripts\") pod \"keystone-7311-account-create-xpsjd\" (UID: \"a11e0baf-fea0-409a-9a16-1a467f83c1bf\") " pod="openstack/keystone-7311-account-create-xpsjd" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.880339 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm66c\" (UniqueName: \"kubernetes.io/projected/a11e0baf-fea0-409a-9a16-1a467f83c1bf-kube-api-access-dm66c\") pod \"keystone-7311-account-create-xpsjd\" (UID: \"a11e0baf-fea0-409a-9a16-1a467f83c1bf\") " pod="openstack/keystone-7311-account-create-xpsjd" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.957194 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-98bd-account-create-x5c52"] Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.958341 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f254c75d-4cc6-43b0-8f88-517f7539d2a5-operator-scripts\") pod \"keystone-db-create-jdldd\" (UID: \"f254c75d-4cc6-43b0-8f88-517f7539d2a5\") " pod="openstack/keystone-db-create-jdldd" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.958584 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb9bd\" (UniqueName: \"kubernetes.io/projected/cd782d01-363c-4e79-9faa-e5266519cf90-kube-api-access-pb9bd\") pod \"placement-db-create-mxkvt\" (UID: \"cd782d01-363c-4e79-9faa-e5266519cf90\") " pod="openstack/placement-db-create-mxkvt" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.958795 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-98bd-account-create-x5c52" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.959182 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f254c75d-4cc6-43b0-8f88-517f7539d2a5-operator-scripts\") pod \"keystone-db-create-jdldd\" (UID: \"f254c75d-4cc6-43b0-8f88-517f7539d2a5\") " pod="openstack/keystone-db-create-jdldd" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.960032 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7311-account-create-xpsjd" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.960343 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd782d01-363c-4e79-9faa-e5266519cf90-operator-scripts\") pod \"placement-db-create-mxkvt\" (UID: \"cd782d01-363c-4e79-9faa-e5266519cf90\") " pod="openstack/placement-db-create-mxkvt" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.960570 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgrh8\" (UniqueName: \"kubernetes.io/projected/f254c75d-4cc6-43b0-8f88-517f7539d2a5-kube-api-access-xgrh8\") pod \"keystone-db-create-jdldd\" (UID: \"f254c75d-4cc6-43b0-8f88-517f7539d2a5\") " pod="openstack/keystone-db-create-jdldd" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.961855 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.971209 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-98bd-account-create-x5c52"] Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.987189 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgrh8\" (UniqueName: \"kubernetes.io/projected/f254c75d-4cc6-43b0-8f88-517f7539d2a5-kube-api-access-xgrh8\") pod \"keystone-db-create-jdldd\" (UID: \"f254c75d-4cc6-43b0-8f88-517f7539d2a5\") " pod="openstack/keystone-db-create-jdldd" Nov 24 17:19:54 crc kubenswrapper[4777]: I1124 17:19:54.991236 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jdldd" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.066064 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/995fd0aa-138b-47b4-a001-3098fcf14e08-operator-scripts\") pod \"placement-98bd-account-create-x5c52\" (UID: \"995fd0aa-138b-47b4-a001-3098fcf14e08\") " pod="openstack/placement-98bd-account-create-x5c52" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.066111 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46cg4\" (UniqueName: \"kubernetes.io/projected/995fd0aa-138b-47b4-a001-3098fcf14e08-kube-api-access-46cg4\") pod \"placement-98bd-account-create-x5c52\" (UID: \"995fd0aa-138b-47b4-a001-3098fcf14e08\") " pod="openstack/placement-98bd-account-create-x5c52" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.066149 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb9bd\" (UniqueName: \"kubernetes.io/projected/cd782d01-363c-4e79-9faa-e5266519cf90-kube-api-access-pb9bd\") pod \"placement-db-create-mxkvt\" (UID: \"cd782d01-363c-4e79-9faa-e5266519cf90\") " pod="openstack/placement-db-create-mxkvt" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.066179 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd782d01-363c-4e79-9faa-e5266519cf90-operator-scripts\") pod \"placement-db-create-mxkvt\" (UID: \"cd782d01-363c-4e79-9faa-e5266519cf90\") " pod="openstack/placement-db-create-mxkvt" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.066847 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd782d01-363c-4e79-9faa-e5266519cf90-operator-scripts\") pod \"placement-db-create-mxkvt\" (UID: \"cd782d01-363c-4e79-9faa-e5266519cf90\") " pod="openstack/placement-db-create-mxkvt" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.072674 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-75wbk"] Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.075333 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-75wbk" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.080178 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-75wbk"] Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.094550 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb9bd\" (UniqueName: \"kubernetes.io/projected/cd782d01-363c-4e79-9faa-e5266519cf90-kube-api-access-pb9bd\") pod \"placement-db-create-mxkvt\" (UID: \"cd782d01-363c-4e79-9faa-e5266519cf90\") " pod="openstack/placement-db-create-mxkvt" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.148279 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mxkvt" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.155414 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-0e91-account-create-jvbdx"] Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.157299 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0e91-account-create-jvbdx" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.159103 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.163105 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-0e91-account-create-jvbdx"] Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.168216 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baaece26-e676-4fcd-a0b2-fd4202d7631a-operator-scripts\") pod \"glance-db-create-75wbk\" (UID: \"baaece26-e676-4fcd-a0b2-fd4202d7631a\") " pod="openstack/glance-db-create-75wbk" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.168378 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/995fd0aa-138b-47b4-a001-3098fcf14e08-operator-scripts\") pod \"placement-98bd-account-create-x5c52\" (UID: \"995fd0aa-138b-47b4-a001-3098fcf14e08\") " pod="openstack/placement-98bd-account-create-x5c52" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.168427 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46cg4\" (UniqueName: \"kubernetes.io/projected/995fd0aa-138b-47b4-a001-3098fcf14e08-kube-api-access-46cg4\") pod \"placement-98bd-account-create-x5c52\" (UID: \"995fd0aa-138b-47b4-a001-3098fcf14e08\") " pod="openstack/placement-98bd-account-create-x5c52" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.168551 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqrht\" (UniqueName: \"kubernetes.io/projected/baaece26-e676-4fcd-a0b2-fd4202d7631a-kube-api-access-jqrht\") pod \"glance-db-create-75wbk\" (UID: \"baaece26-e676-4fcd-a0b2-fd4202d7631a\") " pod="openstack/glance-db-create-75wbk" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.169153 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/995fd0aa-138b-47b4-a001-3098fcf14e08-operator-scripts\") pod \"placement-98bd-account-create-x5c52\" (UID: \"995fd0aa-138b-47b4-a001-3098fcf14e08\") " pod="openstack/placement-98bd-account-create-x5c52" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.189201 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46cg4\" (UniqueName: \"kubernetes.io/projected/995fd0aa-138b-47b4-a001-3098fcf14e08-kube-api-access-46cg4\") pod \"placement-98bd-account-create-x5c52\" (UID: \"995fd0aa-138b-47b4-a001-3098fcf14e08\") " pod="openstack/placement-98bd-account-create-x5c52" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.270521 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqrht\" (UniqueName: \"kubernetes.io/projected/baaece26-e676-4fcd-a0b2-fd4202d7631a-kube-api-access-jqrht\") pod \"glance-db-create-75wbk\" (UID: \"baaece26-e676-4fcd-a0b2-fd4202d7631a\") " pod="openstack/glance-db-create-75wbk" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.270586 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/947652ce-1c76-4d0e-9234-212063ae51c7-operator-scripts\") pod \"glance-0e91-account-create-jvbdx\" (UID: \"947652ce-1c76-4d0e-9234-212063ae51c7\") " pod="openstack/glance-0e91-account-create-jvbdx" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.270672 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzvj6\" (UniqueName: \"kubernetes.io/projected/947652ce-1c76-4d0e-9234-212063ae51c7-kube-api-access-rzvj6\") pod \"glance-0e91-account-create-jvbdx\" (UID: \"947652ce-1c76-4d0e-9234-212063ae51c7\") " pod="openstack/glance-0e91-account-create-jvbdx" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.270742 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baaece26-e676-4fcd-a0b2-fd4202d7631a-operator-scripts\") pod \"glance-db-create-75wbk\" (UID: \"baaece26-e676-4fcd-a0b2-fd4202d7631a\") " pod="openstack/glance-db-create-75wbk" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.271861 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baaece26-e676-4fcd-a0b2-fd4202d7631a-operator-scripts\") pod \"glance-db-create-75wbk\" (UID: \"baaece26-e676-4fcd-a0b2-fd4202d7631a\") " pod="openstack/glance-db-create-75wbk" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.283303 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-98bd-account-create-x5c52" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.286368 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqrht\" (UniqueName: \"kubernetes.io/projected/baaece26-e676-4fcd-a0b2-fd4202d7631a-kube-api-access-jqrht\") pod \"glance-db-create-75wbk\" (UID: \"baaece26-e676-4fcd-a0b2-fd4202d7631a\") " pod="openstack/glance-db-create-75wbk" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.371863 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzvj6\" (UniqueName: \"kubernetes.io/projected/947652ce-1c76-4d0e-9234-212063ae51c7-kube-api-access-rzvj6\") pod \"glance-0e91-account-create-jvbdx\" (UID: \"947652ce-1c76-4d0e-9234-212063ae51c7\") " pod="openstack/glance-0e91-account-create-jvbdx" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.372030 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/947652ce-1c76-4d0e-9234-212063ae51c7-operator-scripts\") pod \"glance-0e91-account-create-jvbdx\" (UID: \"947652ce-1c76-4d0e-9234-212063ae51c7\") " pod="openstack/glance-0e91-account-create-jvbdx" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.372682 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/947652ce-1c76-4d0e-9234-212063ae51c7-operator-scripts\") pod \"glance-0e91-account-create-jvbdx\" (UID: \"947652ce-1c76-4d0e-9234-212063ae51c7\") " pod="openstack/glance-0e91-account-create-jvbdx" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.387647 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzvj6\" (UniqueName: \"kubernetes.io/projected/947652ce-1c76-4d0e-9234-212063ae51c7-kube-api-access-rzvj6\") pod \"glance-0e91-account-create-jvbdx\" (UID: \"947652ce-1c76-4d0e-9234-212063ae51c7\") " pod="openstack/glance-0e91-account-create-jvbdx" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.391386 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-75wbk" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.480161 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0e91-account-create-jvbdx" Nov 24 17:19:55 crc kubenswrapper[4777]: I1124 17:19:55.841138 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f9ab961d-98f3-4a5e-9583-a168fa95f7b4","Type":"ContainerStarted","Data":"e64c61f2703e560ea489d9bd08de334147c4967a65951dfb9b2c81d4873c35dc"} Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.263855 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.493721 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7311-account-create-xpsjd"] Nov 24 17:19:56 crc kubenswrapper[4777]: W1124 17:19:56.495202 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda11e0baf_fea0_409a_9a16_1a467f83c1bf.slice/crio-08595e75fa39b6ddd8334c6ab37268a6f5f471fab10dd1d8e469b10d27194ce8 WatchSource:0}: Error finding container 08595e75fa39b6ddd8334c6ab37268a6f5f471fab10dd1d8e469b10d27194ce8: Status 404 returned error can't find the container with id 08595e75fa39b6ddd8334c6ab37268a6f5f471fab10dd1d8e469b10d27194ce8 Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.523859 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.526032 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.530843 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.531024 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-8stsq" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.531143 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.531918 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.542477 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.559246 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.654984 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-75wbk"] Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.668598 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jdldd"] Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.678921 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-j9kwb"] Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.679197 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" podUID="3370ce09-05a5-4d40-b89b-d75720b39677" containerName="dnsmasq-dns" containerID="cri-o://313143c9d06733f32784b34ade782d8300249aec474b272a14114e2e6ec7ff71" gracePeriod=10 Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.704215 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/001aba35-921b-43b5-87e4-7e92d5595a7b-scripts\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.704302 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/001aba35-921b-43b5-87e4-7e92d5595a7b-config\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.704339 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001aba35-921b-43b5-87e4-7e92d5595a7b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.704378 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/001aba35-921b-43b5-87e4-7e92d5595a7b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.704404 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/001aba35-921b-43b5-87e4-7e92d5595a7b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.704552 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/001aba35-921b-43b5-87e4-7e92d5595a7b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.706183 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnn5z\" (UniqueName: \"kubernetes.io/projected/001aba35-921b-43b5-87e4-7e92d5595a7b-kube-api-access-vnn5z\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.716878 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-98bd-account-create-x5c52"] Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.748029 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-mxkvt"] Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.776152 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-0e91-account-create-jvbdx"] Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.808447 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnn5z\" (UniqueName: \"kubernetes.io/projected/001aba35-921b-43b5-87e4-7e92d5595a7b-kube-api-access-vnn5z\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.813256 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/001aba35-921b-43b5-87e4-7e92d5595a7b-scripts\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.813378 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/001aba35-921b-43b5-87e4-7e92d5595a7b-config\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.813453 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001aba35-921b-43b5-87e4-7e92d5595a7b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.813511 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/001aba35-921b-43b5-87e4-7e92d5595a7b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.813558 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/001aba35-921b-43b5-87e4-7e92d5595a7b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.813618 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/001aba35-921b-43b5-87e4-7e92d5595a7b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.814117 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/001aba35-921b-43b5-87e4-7e92d5595a7b-scripts\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.814118 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/001aba35-921b-43b5-87e4-7e92d5595a7b-config\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.814377 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/001aba35-921b-43b5-87e4-7e92d5595a7b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.820952 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/001aba35-921b-43b5-87e4-7e92d5595a7b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.821624 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001aba35-921b-43b5-87e4-7e92d5595a7b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.832705 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/001aba35-921b-43b5-87e4-7e92d5595a7b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.835958 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnn5z\" (UniqueName: \"kubernetes.io/projected/001aba35-921b-43b5-87e4-7e92d5595a7b-kube-api-access-vnn5z\") pod \"ovn-northd-0\" (UID: \"001aba35-921b-43b5-87e4-7e92d5595a7b\") " pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.859843 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.902417 4777 generic.go:334] "Generic (PLEG): container finished" podID="3370ce09-05a5-4d40-b89b-d75720b39677" containerID="313143c9d06733f32784b34ade782d8300249aec474b272a14114e2e6ec7ff71" exitCode=0 Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.902491 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" event={"ID":"3370ce09-05a5-4d40-b89b-d75720b39677","Type":"ContainerDied","Data":"313143c9d06733f32784b34ade782d8300249aec474b272a14114e2e6ec7ff71"} Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.905941 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7311-account-create-xpsjd" event={"ID":"a11e0baf-fea0-409a-9a16-1a467f83c1bf","Type":"ContainerStarted","Data":"90dfbc6bba00483452e8de630955c422c88de76c8ab125921c52006e19554208"} Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.905978 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7311-account-create-xpsjd" event={"ID":"a11e0baf-fea0-409a-9a16-1a467f83c1bf","Type":"ContainerStarted","Data":"08595e75fa39b6ddd8334c6ab37268a6f5f471fab10dd1d8e469b10d27194ce8"} Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.924631 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"055503f1-ab1c-47db-be36-37a3f572be32","Type":"ContainerStarted","Data":"1b0540e9c949bea450358534dd59351f7d1da40017ade4d66cc96eaa2a3c7fa7"} Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.926367 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mxkvt" event={"ID":"cd782d01-363c-4e79-9faa-e5266519cf90","Type":"ContainerStarted","Data":"62b2696de70d205a9ba6e1ac4bdf7601e8d441468e0478f9fb49f411fbd8ce61"} Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.930858 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-98bd-account-create-x5c52" event={"ID":"995fd0aa-138b-47b4-a001-3098fcf14e08","Type":"ContainerStarted","Data":"da74ac3103edf985c2a2574d064db689fde7e96d0e026e8480a190e92985d953"} Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.934391 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-75wbk" event={"ID":"baaece26-e676-4fcd-a0b2-fd4202d7631a","Type":"ContainerStarted","Data":"278058c35088f0067e41d35986a700a6d334118c05605c226abce8534b59c8fa"} Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.939861 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jdldd" event={"ID":"f254c75d-4cc6-43b0-8f88-517f7539d2a5","Type":"ContainerStarted","Data":"977c7388d2a4601232b152695b05a7a1af254cca355e261b80843d07665914ee"} Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.941850 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0e91-account-create-jvbdx" event={"ID":"947652ce-1c76-4d0e-9234-212063ae51c7","Type":"ContainerStarted","Data":"c53bf1930bce18110ea58952c6d09c6595aedd8648c5775e35d45421dd7f3887"} Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.943322 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4eba3f22-b7ad-464d-a0c9-437c449d124b","Type":"ContainerStarted","Data":"687f2912a7ba6e1f5fc721c17d5d9813047da42feb74dc8b807ec4d3ca2d3660"} Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.962154 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7311-account-create-xpsjd" podStartSLOduration=2.9621216759999998 podStartE2EDuration="2.962121676s" podCreationTimestamp="2025-11-24 17:19:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:19:56.932256579 +0000 UTC m=+1155.091451628" watchObservedRunningTime="2025-11-24 17:19:56.962121676 +0000 UTC m=+1155.121316725" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.967903 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-75wbk" podStartSLOduration=1.9678863519999998 podStartE2EDuration="1.967886352s" podCreationTimestamp="2025-11-24 17:19:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:19:56.953465688 +0000 UTC m=+1155.112660727" watchObservedRunningTime="2025-11-24 17:19:56.967886352 +0000 UTC m=+1155.127081401" Nov 24 17:19:56 crc kubenswrapper[4777]: I1124 17:19:56.975863 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-jdldd" podStartSLOduration=2.97584243 podStartE2EDuration="2.97584243s" podCreationTimestamp="2025-11-24 17:19:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:19:56.969959101 +0000 UTC m=+1155.129154150" watchObservedRunningTime="2025-11-24 17:19:56.97584243 +0000 UTC m=+1155.135037479" Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.438712 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.439716 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=39.318985953 podStartE2EDuration="2m14.439697521s" podCreationTimestamp="2025-11-24 17:17:43 +0000 UTC" firstStartedPulling="2025-11-24 17:17:45.570774706 +0000 UTC m=+1023.729969755" lastFinishedPulling="2025-11-24 17:19:20.691486274 +0000 UTC m=+1118.850681323" observedRunningTime="2025-11-24 17:19:56.99675799 +0000 UTC m=+1155.155953039" watchObservedRunningTime="2025-11-24 17:19:57.439697521 +0000 UTC m=+1155.598892570" Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.442646 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 17:19:57 crc kubenswrapper[4777]: W1124 17:19:57.499795 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod001aba35_921b_43b5_87e4_7e92d5595a7b.slice/crio-35373bb2690ad7e4a0fc04c0a13d751f78792132e21e6429b0e938c9566f7d5d WatchSource:0}: Error finding container 35373bb2690ad7e4a0fc04c0a13d751f78792132e21e6429b0e938c9566f7d5d: Status 404 returned error can't find the container with id 35373bb2690ad7e4a0fc04c0a13d751f78792132e21e6429b0e938c9566f7d5d Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.545508 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ltzh\" (UniqueName: \"kubernetes.io/projected/3370ce09-05a5-4d40-b89b-d75720b39677-kube-api-access-2ltzh\") pod \"3370ce09-05a5-4d40-b89b-d75720b39677\" (UID: \"3370ce09-05a5-4d40-b89b-d75720b39677\") " Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.545653 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3370ce09-05a5-4d40-b89b-d75720b39677-dns-svc\") pod \"3370ce09-05a5-4d40-b89b-d75720b39677\" (UID: \"3370ce09-05a5-4d40-b89b-d75720b39677\") " Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.545780 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3370ce09-05a5-4d40-b89b-d75720b39677-config\") pod \"3370ce09-05a5-4d40-b89b-d75720b39677\" (UID: \"3370ce09-05a5-4d40-b89b-d75720b39677\") " Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.555269 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3370ce09-05a5-4d40-b89b-d75720b39677-kube-api-access-2ltzh" (OuterVolumeSpecName: "kube-api-access-2ltzh") pod "3370ce09-05a5-4d40-b89b-d75720b39677" (UID: "3370ce09-05a5-4d40-b89b-d75720b39677"). InnerVolumeSpecName "kube-api-access-2ltzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.608432 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3370ce09-05a5-4d40-b89b-d75720b39677-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3370ce09-05a5-4d40-b89b-d75720b39677" (UID: "3370ce09-05a5-4d40-b89b-d75720b39677"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.612666 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3370ce09-05a5-4d40-b89b-d75720b39677-config" (OuterVolumeSpecName: "config") pod "3370ce09-05a5-4d40-b89b-d75720b39677" (UID: "3370ce09-05a5-4d40-b89b-d75720b39677"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.647834 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ltzh\" (UniqueName: \"kubernetes.io/projected/3370ce09-05a5-4d40-b89b-d75720b39677-kube-api-access-2ltzh\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.647864 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3370ce09-05a5-4d40-b89b-d75720b39677-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.647873 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3370ce09-05a5-4d40-b89b-d75720b39677-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.959512 4777 generic.go:334] "Generic (PLEG): container finished" podID="995fd0aa-138b-47b4-a001-3098fcf14e08" containerID="cdc32b64f9d9589e5cf0fc23fbaa454e1bced455393d8456e88baab88c5f3f2c" exitCode=0 Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.960403 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-98bd-account-create-x5c52" event={"ID":"995fd0aa-138b-47b4-a001-3098fcf14e08","Type":"ContainerDied","Data":"cdc32b64f9d9589e5cf0fc23fbaa454e1bced455393d8456e88baab88c5f3f2c"} Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.962818 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" event={"ID":"3370ce09-05a5-4d40-b89b-d75720b39677","Type":"ContainerDied","Data":"8005b69fc25998703c4c758ebf5d03c030e9a6be084e1edf3e5d2cb6a9713ba9"} Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.962865 4777 scope.go:117] "RemoveContainer" containerID="313143c9d06733f32784b34ade782d8300249aec474b272a14114e2e6ec7ff71" Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.963032 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-j9kwb" Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.965243 4777 generic.go:334] "Generic (PLEG): container finished" podID="baaece26-e676-4fcd-a0b2-fd4202d7631a" containerID="0fc4f625856d9a7ad250bfd9e7804dd80ee0c0dacae165c7e60b289b70473a20" exitCode=0 Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.965495 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-75wbk" event={"ID":"baaece26-e676-4fcd-a0b2-fd4202d7631a","Type":"ContainerDied","Data":"0fc4f625856d9a7ad250bfd9e7804dd80ee0c0dacae165c7e60b289b70473a20"} Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.967654 4777 generic.go:334] "Generic (PLEG): container finished" podID="a11e0baf-fea0-409a-9a16-1a467f83c1bf" containerID="90dfbc6bba00483452e8de630955c422c88de76c8ab125921c52006e19554208" exitCode=0 Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.967697 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7311-account-create-xpsjd" event={"ID":"a11e0baf-fea0-409a-9a16-1a467f83c1bf","Type":"ContainerDied","Data":"90dfbc6bba00483452e8de630955c422c88de76c8ab125921c52006e19554208"} Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.968866 4777 generic.go:334] "Generic (PLEG): container finished" podID="f254c75d-4cc6-43b0-8f88-517f7539d2a5" containerID="16914b569eb76eb8eec42c98f9b054e6f10324333886c0bba58e9c4445b0e1f6" exitCode=0 Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.968900 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jdldd" event={"ID":"f254c75d-4cc6-43b0-8f88-517f7539d2a5","Type":"ContainerDied","Data":"16914b569eb76eb8eec42c98f9b054e6f10324333886c0bba58e9c4445b0e1f6"} Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.970103 4777 generic.go:334] "Generic (PLEG): container finished" podID="cd782d01-363c-4e79-9faa-e5266519cf90" containerID="c493219defb7ac0b49f6571834ff1d085dd31a86ddbfc2716ef1d14a29eb16b0" exitCode=0 Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.970143 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mxkvt" event={"ID":"cd782d01-363c-4e79-9faa-e5266519cf90","Type":"ContainerDied","Data":"c493219defb7ac0b49f6571834ff1d085dd31a86ddbfc2716ef1d14a29eb16b0"} Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.985552 4777 generic.go:334] "Generic (PLEG): container finished" podID="947652ce-1c76-4d0e-9234-212063ae51c7" containerID="93cccd55686027e03ab1720528be46272682e0d0b294a31e5f9fe1534463b6c7" exitCode=0 Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.985706 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0e91-account-create-jvbdx" event={"ID":"947652ce-1c76-4d0e-9234-212063ae51c7","Type":"ContainerDied","Data":"93cccd55686027e03ab1720528be46272682e0d0b294a31e5f9fe1534463b6c7"} Nov 24 17:19:57 crc kubenswrapper[4777]: I1124 17:19:57.988211 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"001aba35-921b-43b5-87e4-7e92d5595a7b","Type":"ContainerStarted","Data":"35373bb2690ad7e4a0fc04c0a13d751f78792132e21e6429b0e938c9566f7d5d"} Nov 24 17:19:58 crc kubenswrapper[4777]: I1124 17:19:58.000372 4777 scope.go:117] "RemoveContainer" containerID="6b091a1321d3320dd38a52f6501db9464a59367e9933462a9fdb1c64f7300ce2" Nov 24 17:19:58 crc kubenswrapper[4777]: I1124 17:19:58.109299 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-j9kwb"] Nov 24 17:19:58 crc kubenswrapper[4777]: I1124 17:19:58.120570 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-j9kwb"] Nov 24 17:19:58 crc kubenswrapper[4777]: I1124 17:19:58.724030 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="f68bfbdc-e400-488d-9f32-ea90fd1160e3" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 17:19:58 crc kubenswrapper[4777]: I1124 17:19:58.999322 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"055503f1-ab1c-47db-be36-37a3f572be32","Type":"ContainerStarted","Data":"89001a2074023ef18199bc038925f9aeee2c5204448cae557fdc1887f5600516"} Nov 24 17:19:58 crc kubenswrapper[4777]: I1124 17:19:58.999372 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Nov 24 17:19:59 crc kubenswrapper[4777]: I1124 17:19:59.032857 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=83.641769568 podStartE2EDuration="2m13.032839178s" podCreationTimestamp="2025-11-24 17:17:46 +0000 UTC" firstStartedPulling="2025-11-24 17:19:06.643853599 +0000 UTC m=+1104.803048648" lastFinishedPulling="2025-11-24 17:19:56.034923199 +0000 UTC m=+1154.194118258" observedRunningTime="2025-11-24 17:19:59.026491325 +0000 UTC m=+1157.185686394" watchObservedRunningTime="2025-11-24 17:19:59.032839178 +0000 UTC m=+1157.192034227" Nov 24 17:19:59 crc kubenswrapper[4777]: I1124 17:19:59.267915 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3370ce09-05a5-4d40-b89b-d75720b39677" path="/var/lib/kubelet/pods/3370ce09-05a5-4d40-b89b-d75720b39677/volumes" Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.016459 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.599316 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-98bd-account-create-x5c52" Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.609825 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mxkvt" Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.717613 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-q4l2k" podUID="73928e73-667c-4b69-aed9-72f32012fbdc" containerName="ovn-controller" probeResult="failure" output=< Nov 24 17:20:00 crc kubenswrapper[4777]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 17:20:00 crc kubenswrapper[4777]: > Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.775705 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb9bd\" (UniqueName: \"kubernetes.io/projected/cd782d01-363c-4e79-9faa-e5266519cf90-kube-api-access-pb9bd\") pod \"cd782d01-363c-4e79-9faa-e5266519cf90\" (UID: \"cd782d01-363c-4e79-9faa-e5266519cf90\") " Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.775749 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46cg4\" (UniqueName: \"kubernetes.io/projected/995fd0aa-138b-47b4-a001-3098fcf14e08-kube-api-access-46cg4\") pod \"995fd0aa-138b-47b4-a001-3098fcf14e08\" (UID: \"995fd0aa-138b-47b4-a001-3098fcf14e08\") " Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.775918 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/995fd0aa-138b-47b4-a001-3098fcf14e08-operator-scripts\") pod \"995fd0aa-138b-47b4-a001-3098fcf14e08\" (UID: \"995fd0aa-138b-47b4-a001-3098fcf14e08\") " Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.776058 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd782d01-363c-4e79-9faa-e5266519cf90-operator-scripts\") pod \"cd782d01-363c-4e79-9faa-e5266519cf90\" (UID: \"cd782d01-363c-4e79-9faa-e5266519cf90\") " Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.777158 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd782d01-363c-4e79-9faa-e5266519cf90-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cd782d01-363c-4e79-9faa-e5266519cf90" (UID: "cd782d01-363c-4e79-9faa-e5266519cf90"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.777271 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/995fd0aa-138b-47b4-a001-3098fcf14e08-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "995fd0aa-138b-47b4-a001-3098fcf14e08" (UID: "995fd0aa-138b-47b4-a001-3098fcf14e08"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.782150 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd782d01-363c-4e79-9faa-e5266519cf90-kube-api-access-pb9bd" (OuterVolumeSpecName: "kube-api-access-pb9bd") pod "cd782d01-363c-4e79-9faa-e5266519cf90" (UID: "cd782d01-363c-4e79-9faa-e5266519cf90"). InnerVolumeSpecName "kube-api-access-pb9bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.782246 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/995fd0aa-138b-47b4-a001-3098fcf14e08-kube-api-access-46cg4" (OuterVolumeSpecName: "kube-api-access-46cg4") pod "995fd0aa-138b-47b4-a001-3098fcf14e08" (UID: "995fd0aa-138b-47b4-a001-3098fcf14e08"). InnerVolumeSpecName "kube-api-access-46cg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.877481 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/995fd0aa-138b-47b4-a001-3098fcf14e08-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.877510 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd782d01-363c-4e79-9faa-e5266519cf90-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.877520 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb9bd\" (UniqueName: \"kubernetes.io/projected/cd782d01-363c-4e79-9faa-e5266519cf90-kube-api-access-pb9bd\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.877530 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46cg4\" (UniqueName: \"kubernetes.io/projected/995fd0aa-138b-47b4-a001-3098fcf14e08-kube-api-access-46cg4\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.955436 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0e91-account-create-jvbdx" Nov 24 17:20:00 crc kubenswrapper[4777]: I1124 17:20:00.972138 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-75wbk" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.001425 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7311-account-create-xpsjd" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.006382 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jdldd" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.032850 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jdldd" event={"ID":"f254c75d-4cc6-43b0-8f88-517f7539d2a5","Type":"ContainerDied","Data":"977c7388d2a4601232b152695b05a7a1af254cca355e261b80843d07665914ee"} Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.032937 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="977c7388d2a4601232b152695b05a7a1af254cca355e261b80843d07665914ee" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.033021 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jdldd" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.037041 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mxkvt" event={"ID":"cd782d01-363c-4e79-9faa-e5266519cf90","Type":"ContainerDied","Data":"62b2696de70d205a9ba6e1ac4bdf7601e8d441468e0478f9fb49f411fbd8ce61"} Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.037074 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62b2696de70d205a9ba6e1ac4bdf7601e8d441468e0478f9fb49f411fbd8ce61" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.039911 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mxkvt" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.046063 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0e91-account-create-jvbdx" event={"ID":"947652ce-1c76-4d0e-9234-212063ae51c7","Type":"ContainerDied","Data":"c53bf1930bce18110ea58952c6d09c6595aedd8648c5775e35d45421dd7f3887"} Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.046098 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c53bf1930bce18110ea58952c6d09c6595aedd8648c5775e35d45421dd7f3887" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.046147 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0e91-account-create-jvbdx" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.048214 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-98bd-account-create-x5c52" event={"ID":"995fd0aa-138b-47b4-a001-3098fcf14e08","Type":"ContainerDied","Data":"da74ac3103edf985c2a2574d064db689fde7e96d0e026e8480a190e92985d953"} Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.048238 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da74ac3103edf985c2a2574d064db689fde7e96d0e026e8480a190e92985d953" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.048276 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-98bd-account-create-x5c52" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.050821 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-75wbk" event={"ID":"baaece26-e676-4fcd-a0b2-fd4202d7631a","Type":"ContainerDied","Data":"278058c35088f0067e41d35986a700a6d334118c05605c226abce8534b59c8fa"} Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.050875 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="278058c35088f0067e41d35986a700a6d334118c05605c226abce8534b59c8fa" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.050990 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-75wbk" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.060124 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7311-account-create-xpsjd" event={"ID":"a11e0baf-fea0-409a-9a16-1a467f83c1bf","Type":"ContainerDied","Data":"08595e75fa39b6ddd8334c6ab37268a6f5f471fab10dd1d8e469b10d27194ce8"} Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.060178 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08595e75fa39b6ddd8334c6ab37268a6f5f471fab10dd1d8e469b10d27194ce8" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.061456 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7311-account-create-xpsjd" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.079597 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqrht\" (UniqueName: \"kubernetes.io/projected/baaece26-e676-4fcd-a0b2-fd4202d7631a-kube-api-access-jqrht\") pod \"baaece26-e676-4fcd-a0b2-fd4202d7631a\" (UID: \"baaece26-e676-4fcd-a0b2-fd4202d7631a\") " Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.079714 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/947652ce-1c76-4d0e-9234-212063ae51c7-operator-scripts\") pod \"947652ce-1c76-4d0e-9234-212063ae51c7\" (UID: \"947652ce-1c76-4d0e-9234-212063ae51c7\") " Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.079763 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzvj6\" (UniqueName: \"kubernetes.io/projected/947652ce-1c76-4d0e-9234-212063ae51c7-kube-api-access-rzvj6\") pod \"947652ce-1c76-4d0e-9234-212063ae51c7\" (UID: \"947652ce-1c76-4d0e-9234-212063ae51c7\") " Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.079825 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baaece26-e676-4fcd-a0b2-fd4202d7631a-operator-scripts\") pod \"baaece26-e676-4fcd-a0b2-fd4202d7631a\" (UID: \"baaece26-e676-4fcd-a0b2-fd4202d7631a\") " Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.081739 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baaece26-e676-4fcd-a0b2-fd4202d7631a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "baaece26-e676-4fcd-a0b2-fd4202d7631a" (UID: "baaece26-e676-4fcd-a0b2-fd4202d7631a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.081814 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/947652ce-1c76-4d0e-9234-212063ae51c7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "947652ce-1c76-4d0e-9234-212063ae51c7" (UID: "947652ce-1c76-4d0e-9234-212063ae51c7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.083096 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baaece26-e676-4fcd-a0b2-fd4202d7631a-kube-api-access-jqrht" (OuterVolumeSpecName: "kube-api-access-jqrht") pod "baaece26-e676-4fcd-a0b2-fd4202d7631a" (UID: "baaece26-e676-4fcd-a0b2-fd4202d7631a"). InnerVolumeSpecName "kube-api-access-jqrht". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.083624 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/947652ce-1c76-4d0e-9234-212063ae51c7-kube-api-access-rzvj6" (OuterVolumeSpecName: "kube-api-access-rzvj6") pod "947652ce-1c76-4d0e-9234-212063ae51c7" (UID: "947652ce-1c76-4d0e-9234-212063ae51c7"). InnerVolumeSpecName "kube-api-access-rzvj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.181012 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a11e0baf-fea0-409a-9a16-1a467f83c1bf-operator-scripts\") pod \"a11e0baf-fea0-409a-9a16-1a467f83c1bf\" (UID: \"a11e0baf-fea0-409a-9a16-1a467f83c1bf\") " Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.181078 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f254c75d-4cc6-43b0-8f88-517f7539d2a5-operator-scripts\") pod \"f254c75d-4cc6-43b0-8f88-517f7539d2a5\" (UID: \"f254c75d-4cc6-43b0-8f88-517f7539d2a5\") " Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.181102 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgrh8\" (UniqueName: \"kubernetes.io/projected/f254c75d-4cc6-43b0-8f88-517f7539d2a5-kube-api-access-xgrh8\") pod \"f254c75d-4cc6-43b0-8f88-517f7539d2a5\" (UID: \"f254c75d-4cc6-43b0-8f88-517f7539d2a5\") " Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.181134 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dm66c\" (UniqueName: \"kubernetes.io/projected/a11e0baf-fea0-409a-9a16-1a467f83c1bf-kube-api-access-dm66c\") pod \"a11e0baf-fea0-409a-9a16-1a467f83c1bf\" (UID: \"a11e0baf-fea0-409a-9a16-1a467f83c1bf\") " Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.181543 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqrht\" (UniqueName: \"kubernetes.io/projected/baaece26-e676-4fcd-a0b2-fd4202d7631a-kube-api-access-jqrht\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.181563 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/947652ce-1c76-4d0e-9234-212063ae51c7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.181572 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzvj6\" (UniqueName: \"kubernetes.io/projected/947652ce-1c76-4d0e-9234-212063ae51c7-kube-api-access-rzvj6\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.181581 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baaece26-e676-4fcd-a0b2-fd4202d7631a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.182644 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f254c75d-4cc6-43b0-8f88-517f7539d2a5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f254c75d-4cc6-43b0-8f88-517f7539d2a5" (UID: "f254c75d-4cc6-43b0-8f88-517f7539d2a5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.184500 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a11e0baf-fea0-409a-9a16-1a467f83c1bf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a11e0baf-fea0-409a-9a16-1a467f83c1bf" (UID: "a11e0baf-fea0-409a-9a16-1a467f83c1bf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.189259 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a11e0baf-fea0-409a-9a16-1a467f83c1bf-kube-api-access-dm66c" (OuterVolumeSpecName: "kube-api-access-dm66c") pod "a11e0baf-fea0-409a-9a16-1a467f83c1bf" (UID: "a11e0baf-fea0-409a-9a16-1a467f83c1bf"). InnerVolumeSpecName "kube-api-access-dm66c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.193660 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f254c75d-4cc6-43b0-8f88-517f7539d2a5-kube-api-access-xgrh8" (OuterVolumeSpecName: "kube-api-access-xgrh8") pod "f254c75d-4cc6-43b0-8f88-517f7539d2a5" (UID: "f254c75d-4cc6-43b0-8f88-517f7539d2a5"). InnerVolumeSpecName "kube-api-access-xgrh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.283122 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a11e0baf-fea0-409a-9a16-1a467f83c1bf-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.283164 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f254c75d-4cc6-43b0-8f88-517f7539d2a5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.283177 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgrh8\" (UniqueName: \"kubernetes.io/projected/f254c75d-4cc6-43b0-8f88-517f7539d2a5-kube-api-access-xgrh8\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:01 crc kubenswrapper[4777]: I1124 17:20:01.283193 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dm66c\" (UniqueName: \"kubernetes.io/projected/a11e0baf-fea0-409a-9a16-1a467f83c1bf-kube-api-access-dm66c\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:02 crc kubenswrapper[4777]: I1124 17:20:02.082597 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xgm6s" event={"ID":"3ec24d0c-556b-4a24-8f42-38fed08efe6b","Type":"ContainerStarted","Data":"62ffc2935295b2dbd24c5488e3bdb6d73839009626d5cf6cd3810e31fc0aed49"} Nov 24 17:20:02 crc kubenswrapper[4777]: I1124 17:20:02.086752 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f9ab961d-98f3-4a5e-9583-a168fa95f7b4","Type":"ContainerStarted","Data":"05e21734aec05ff774fe535fea9c31c3c6dafffa7b3c7ac51f401de3cd19fbba"} Nov 24 17:20:02 crc kubenswrapper[4777]: I1124 17:20:02.090795 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"001aba35-921b-43b5-87e4-7e92d5595a7b","Type":"ContainerStarted","Data":"183fe4a9065426da10de3ed0684d69fe21b34801ae4c8773b618a92c5852767e"} Nov 24 17:20:02 crc kubenswrapper[4777]: I1124 17:20:02.090839 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"001aba35-921b-43b5-87e4-7e92d5595a7b","Type":"ContainerStarted","Data":"9ecfdd48e2c60382544afb67587779d3d1740c819b89879cc56fef024d8413a2"} Nov 24 17:20:02 crc kubenswrapper[4777]: I1124 17:20:02.090977 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 24 17:20:02 crc kubenswrapper[4777]: I1124 17:20:02.121868 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-xgm6s" podStartSLOduration=4.212528162 podStartE2EDuration="44.121845391s" podCreationTimestamp="2025-11-24 17:19:18 +0000 UTC" firstStartedPulling="2025-11-24 17:19:21.155408437 +0000 UTC m=+1119.314603486" lastFinishedPulling="2025-11-24 17:20:01.064725666 +0000 UTC m=+1159.223920715" observedRunningTime="2025-11-24 17:20:02.105707317 +0000 UTC m=+1160.264902356" watchObservedRunningTime="2025-11-24 17:20:02.121845391 +0000 UTC m=+1160.281040450" Nov 24 17:20:02 crc kubenswrapper[4777]: I1124 17:20:02.143849 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.590339439 podStartE2EDuration="6.143828781s" podCreationTimestamp="2025-11-24 17:19:56 +0000 UTC" firstStartedPulling="2025-11-24 17:19:57.508801544 +0000 UTC m=+1155.667996593" lastFinishedPulling="2025-11-24 17:20:01.062290856 +0000 UTC m=+1159.221485935" observedRunningTime="2025-11-24 17:20:02.126491234 +0000 UTC m=+1160.285686283" watchObservedRunningTime="2025-11-24 17:20:02.143828781 +0000 UTC m=+1160.303023850" Nov 24 17:20:02 crc kubenswrapper[4777]: I1124 17:20:02.152460 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=81.248534541 podStartE2EDuration="2m16.152442369s" podCreationTimestamp="2025-11-24 17:17:46 +0000 UTC" firstStartedPulling="2025-11-24 17:19:06.093507256 +0000 UTC m=+1104.252702305" lastFinishedPulling="2025-11-24 17:20:00.997415084 +0000 UTC m=+1159.156610133" observedRunningTime="2025-11-24 17:20:02.149931337 +0000 UTC m=+1160.309126406" watchObservedRunningTime="2025-11-24 17:20:02.152442369 +0000 UTC m=+1160.311637428" Nov 24 17:20:02 crc kubenswrapper[4777]: I1124 17:20:02.750822 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:02 crc kubenswrapper[4777]: I1124 17:20:02.750907 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:02 crc kubenswrapper[4777]: I1124 17:20:02.753057 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:03 crc kubenswrapper[4777]: I1124 17:20:03.112224 4777 generic.go:334] "Generic (PLEG): container finished" podID="7163366e-9a39-4379-af0a-d6c7c998ab7f" containerID="51d916dedea9f0bddc55f0a8a97bf14ba1e65ae0a91b751dee93cb87d18481b8" exitCode=0 Nov 24 17:20:03 crc kubenswrapper[4777]: I1124 17:20:03.112295 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7163366e-9a39-4379-af0a-d6c7c998ab7f","Type":"ContainerDied","Data":"51d916dedea9f0bddc55f0a8a97bf14ba1e65ae0a91b751dee93cb87d18481b8"} Nov 24 17:20:03 crc kubenswrapper[4777]: I1124 17:20:03.121925 4777 generic.go:334] "Generic (PLEG): container finished" podID="4c707277-46e4-42ea-9076-566b2d01f544" containerID="9dfcd853b3fc5572da0af858e5836f1849e0716f5fca857126f422e455bcc6b7" exitCode=0 Nov 24 17:20:03 crc kubenswrapper[4777]: I1124 17:20:03.122012 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4c707277-46e4-42ea-9076-566b2d01f544","Type":"ContainerDied","Data":"9dfcd853b3fc5572da0af858e5836f1849e0716f5fca857126f422e455bcc6b7"} Nov 24 17:20:03 crc kubenswrapper[4777]: I1124 17:20:03.123449 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:04 crc kubenswrapper[4777]: I1124 17:20:04.134274 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4c707277-46e4-42ea-9076-566b2d01f544","Type":"ContainerStarted","Data":"e5a70124332b963c74c5101f8eebeb8e795be45b3dcb774cd21bae785822c37b"} Nov 24 17:20:04 crc kubenswrapper[4777]: I1124 17:20:04.134955 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 17:20:04 crc kubenswrapper[4777]: I1124 17:20:04.137766 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7163366e-9a39-4379-af0a-d6c7c998ab7f","Type":"ContainerStarted","Data":"de911e01738284bd59a6b0afed7357ce4a53fef2f11ed85c0d07b82a41044436"} Nov 24 17:20:04 crc kubenswrapper[4777]: I1124 17:20:04.138258 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:20:04 crc kubenswrapper[4777]: I1124 17:20:04.160593 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=46.74554328 podStartE2EDuration="2m25.160572525s" podCreationTimestamp="2025-11-24 17:17:39 +0000 UTC" firstStartedPulling="2025-11-24 17:17:41.904817064 +0000 UTC m=+1020.064012123" lastFinishedPulling="2025-11-24 17:19:20.319846319 +0000 UTC m=+1118.479041368" observedRunningTime="2025-11-24 17:20:04.1548118 +0000 UTC m=+1162.314006859" watchObservedRunningTime="2025-11-24 17:20:04.160572525 +0000 UTC m=+1162.319767574" Nov 24 17:20:04 crc kubenswrapper[4777]: I1124 17:20:04.188073 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=45.55020678 podStartE2EDuration="2m24.188056414s" podCreationTimestamp="2025-11-24 17:17:40 +0000 UTC" firstStartedPulling="2025-11-24 17:17:42.05467836 +0000 UTC m=+1020.213873409" lastFinishedPulling="2025-11-24 17:19:20.692527984 +0000 UTC m=+1118.851723043" observedRunningTime="2025-11-24 17:20:04.184524212 +0000 UTC m=+1162.343719271" watchObservedRunningTime="2025-11-24 17:20:04.188056414 +0000 UTC m=+1162.347251483" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.119287 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.119332 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.209675 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.300517 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.383304 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-swrgj"] Nov 24 17:20:05 crc kubenswrapper[4777]: E1124 17:20:05.383886 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baaece26-e676-4fcd-a0b2-fd4202d7631a" containerName="mariadb-database-create" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.383903 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="baaece26-e676-4fcd-a0b2-fd4202d7631a" containerName="mariadb-database-create" Nov 24 17:20:05 crc kubenswrapper[4777]: E1124 17:20:05.383923 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a11e0baf-fea0-409a-9a16-1a467f83c1bf" containerName="mariadb-account-create" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.383930 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="a11e0baf-fea0-409a-9a16-1a467f83c1bf" containerName="mariadb-account-create" Nov 24 17:20:05 crc kubenswrapper[4777]: E1124 17:20:05.383937 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd782d01-363c-4e79-9faa-e5266519cf90" containerName="mariadb-database-create" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.383943 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd782d01-363c-4e79-9faa-e5266519cf90" containerName="mariadb-database-create" Nov 24 17:20:05 crc kubenswrapper[4777]: E1124 17:20:05.383953 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f254c75d-4cc6-43b0-8f88-517f7539d2a5" containerName="mariadb-database-create" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.383959 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="f254c75d-4cc6-43b0-8f88-517f7539d2a5" containerName="mariadb-database-create" Nov 24 17:20:05 crc kubenswrapper[4777]: E1124 17:20:05.383983 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995fd0aa-138b-47b4-a001-3098fcf14e08" containerName="mariadb-account-create" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.383989 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="995fd0aa-138b-47b4-a001-3098fcf14e08" containerName="mariadb-account-create" Nov 24 17:20:05 crc kubenswrapper[4777]: E1124 17:20:05.384000 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3370ce09-05a5-4d40-b89b-d75720b39677" containerName="dnsmasq-dns" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.384006 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="3370ce09-05a5-4d40-b89b-d75720b39677" containerName="dnsmasq-dns" Nov 24 17:20:05 crc kubenswrapper[4777]: E1124 17:20:05.384025 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3370ce09-05a5-4d40-b89b-d75720b39677" containerName="init" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.384032 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="3370ce09-05a5-4d40-b89b-d75720b39677" containerName="init" Nov 24 17:20:05 crc kubenswrapper[4777]: E1124 17:20:05.384045 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="947652ce-1c76-4d0e-9234-212063ae51c7" containerName="mariadb-account-create" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.384051 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="947652ce-1c76-4d0e-9234-212063ae51c7" containerName="mariadb-account-create" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.384205 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="947652ce-1c76-4d0e-9234-212063ae51c7" containerName="mariadb-account-create" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.384214 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="995fd0aa-138b-47b4-a001-3098fcf14e08" containerName="mariadb-account-create" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.384222 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="a11e0baf-fea0-409a-9a16-1a467f83c1bf" containerName="mariadb-account-create" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.384238 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="f254c75d-4cc6-43b0-8f88-517f7539d2a5" containerName="mariadb-database-create" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.384247 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd782d01-363c-4e79-9faa-e5266519cf90" containerName="mariadb-database-create" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.384258 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="baaece26-e676-4fcd-a0b2-fd4202d7631a" containerName="mariadb-database-create" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.384268 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="3370ce09-05a5-4d40-b89b-d75720b39677" containerName="dnsmasq-dns" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.384875 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.388005 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.388218 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-v86dz" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.443931 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-swrgj"] Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.475267 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-db-sync-config-data\") pod \"glance-db-sync-swrgj\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.475329 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmwh6\" (UniqueName: \"kubernetes.io/projected/7b60b793-78a5-406c-9ed6-e582dfba568b-kube-api-access-tmwh6\") pod \"glance-db-sync-swrgj\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.475406 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-config-data\") pod \"glance-db-sync-swrgj\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.475480 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-combined-ca-bundle\") pod \"glance-db-sync-swrgj\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.577247 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-combined-ca-bundle\") pod \"glance-db-sync-swrgj\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.577386 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-db-sync-config-data\") pod \"glance-db-sync-swrgj\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.577412 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmwh6\" (UniqueName: \"kubernetes.io/projected/7b60b793-78a5-406c-9ed6-e582dfba568b-kube-api-access-tmwh6\") pod \"glance-db-sync-swrgj\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.577451 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-config-data\") pod \"glance-db-sync-swrgj\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.584085 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-db-sync-config-data\") pod \"glance-db-sync-swrgj\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.584294 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-combined-ca-bundle\") pod \"glance-db-sync-swrgj\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.584599 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-config-data\") pod \"glance-db-sync-swrgj\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.595686 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmwh6\" (UniqueName: \"kubernetes.io/projected/7b60b793-78a5-406c-9ed6-e582dfba568b-kube-api-access-tmwh6\") pod \"glance-db-sync-swrgj\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.699816 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.700091 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerName="prometheus" containerID="cri-o://8bc55352e58db3657ba1188879f371842295e1d4d4f5233b6c23347da8f5c05e" gracePeriod=600 Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.700148 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerName="thanos-sidecar" containerID="cri-o://05e21734aec05ff774fe535fea9c31c3c6dafffa7b3c7ac51f401de3cd19fbba" gracePeriod=600 Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.700201 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerName="config-reloader" containerID="cri-o://e64c61f2703e560ea489d9bd08de334147c4967a65951dfb9b2c81d4873c35dc" gracePeriod=600 Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.701802 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:05 crc kubenswrapper[4777]: I1124 17:20:05.715447 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-q4l2k" podUID="73928e73-667c-4b69-aed9-72f32012fbdc" containerName="ovn-controller" probeResult="failure" output=< Nov 24 17:20:05 crc kubenswrapper[4777]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 17:20:05 crc kubenswrapper[4777]: > Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.173401 4777 generic.go:334] "Generic (PLEG): container finished" podID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerID="05e21734aec05ff774fe535fea9c31c3c6dafffa7b3c7ac51f401de3cd19fbba" exitCode=0 Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.173439 4777 generic.go:334] "Generic (PLEG): container finished" podID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerID="e64c61f2703e560ea489d9bd08de334147c4967a65951dfb9b2c81d4873c35dc" exitCode=0 Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.173449 4777 generic.go:334] "Generic (PLEG): container finished" podID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerID="8bc55352e58db3657ba1188879f371842295e1d4d4f5233b6c23347da8f5c05e" exitCode=0 Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.173482 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f9ab961d-98f3-4a5e-9583-a168fa95f7b4","Type":"ContainerDied","Data":"05e21734aec05ff774fe535fea9c31c3c6dafffa7b3c7ac51f401de3cd19fbba"} Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.173524 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f9ab961d-98f3-4a5e-9583-a168fa95f7b4","Type":"ContainerDied","Data":"e64c61f2703e560ea489d9bd08de334147c4967a65951dfb9b2c81d4873c35dc"} Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.173536 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f9ab961d-98f3-4a5e-9583-a168fa95f7b4","Type":"ContainerDied","Data":"8bc55352e58db3657ba1188879f371842295e1d4d4f5233b6c23347da8f5c05e"} Nov 24 17:20:06 crc kubenswrapper[4777]: W1124 17:20:06.447118 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b60b793_78a5_406c_9ed6_e582dfba568b.slice/crio-f827e21383ff8622f7cfba44697eb036fa2556d1b57b579fc0bc0dcd63d10c1c WatchSource:0}: Error finding container f827e21383ff8622f7cfba44697eb036fa2556d1b57b579fc0bc0dcd63d10c1c: Status 404 returned error can't find the container with id f827e21383ff8622f7cfba44697eb036fa2556d1b57b579fc0bc0dcd63d10c1c Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.449479 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-swrgj"] Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.734343 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.895202 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-web-config\") pod \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.895387 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-872360a4-14ac-4955-839d-37d8c1b22667\") pod \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.895439 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-config-out\") pod \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.895473 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-prometheus-metric-storage-rulefiles-0\") pod \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.895506 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjm9x\" (UniqueName: \"kubernetes.io/projected/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-kube-api-access-gjm9x\") pod \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.895538 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-config\") pod \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.895609 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-thanos-prometheus-http-client-file\") pod \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.895685 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-tls-assets\") pod \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\" (UID: \"f9ab961d-98f3-4a5e-9583-a168fa95f7b4\") " Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.897403 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "f9ab961d-98f3-4a5e-9583-a168fa95f7b4" (UID: "f9ab961d-98f3-4a5e-9583-a168fa95f7b4"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.902696 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-config-out" (OuterVolumeSpecName: "config-out") pod "f9ab961d-98f3-4a5e-9583-a168fa95f7b4" (UID: "f9ab961d-98f3-4a5e-9583-a168fa95f7b4"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.903067 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "f9ab961d-98f3-4a5e-9583-a168fa95f7b4" (UID: "f9ab961d-98f3-4a5e-9583-a168fa95f7b4"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.903808 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-config" (OuterVolumeSpecName: "config") pod "f9ab961d-98f3-4a5e-9583-a168fa95f7b4" (UID: "f9ab961d-98f3-4a5e-9583-a168fa95f7b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.905129 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "f9ab961d-98f3-4a5e-9583-a168fa95f7b4" (UID: "f9ab961d-98f3-4a5e-9583-a168fa95f7b4"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.909657 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-kube-api-access-gjm9x" (OuterVolumeSpecName: "kube-api-access-gjm9x") pod "f9ab961d-98f3-4a5e-9583-a168fa95f7b4" (UID: "f9ab961d-98f3-4a5e-9583-a168fa95f7b4"). InnerVolumeSpecName "kube-api-access-gjm9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.921158 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-872360a4-14ac-4955-839d-37d8c1b22667" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "f9ab961d-98f3-4a5e-9583-a168fa95f7b4" (UID: "f9ab961d-98f3-4a5e-9583-a168fa95f7b4"). InnerVolumeSpecName "pvc-872360a4-14ac-4955-839d-37d8c1b22667". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.941803 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-web-config" (OuterVolumeSpecName: "web-config") pod "f9ab961d-98f3-4a5e-9583-a168fa95f7b4" (UID: "f9ab961d-98f3-4a5e-9583-a168fa95f7b4"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.997119 4777 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-config-out\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.997149 4777 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.997160 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjm9x\" (UniqueName: \"kubernetes.io/projected/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-kube-api-access-gjm9x\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.997170 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.997179 4777 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.997188 4777 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-tls-assets\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.997198 4777 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f9ab961d-98f3-4a5e-9583-a168fa95f7b4-web-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:06 crc kubenswrapper[4777]: I1124 17:20:06.997227 4777 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-872360a4-14ac-4955-839d-37d8c1b22667\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-872360a4-14ac-4955-839d-37d8c1b22667\") on node \"crc\" " Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.018700 4777 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.018893 4777 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-872360a4-14ac-4955-839d-37d8c1b22667" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-872360a4-14ac-4955-839d-37d8c1b22667") on node "crc" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.100094 4777 reconciler_common.go:293] "Volume detached for volume \"pvc-872360a4-14ac-4955-839d-37d8c1b22667\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-872360a4-14ac-4955-839d-37d8c1b22667\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.192776 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-swrgj" event={"ID":"7b60b793-78a5-406c-9ed6-e582dfba568b","Type":"ContainerStarted","Data":"f827e21383ff8622f7cfba44697eb036fa2556d1b57b579fc0bc0dcd63d10c1c"} Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.201359 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f9ab961d-98f3-4a5e-9583-a168fa95f7b4","Type":"ContainerDied","Data":"2771454974f38c8e15ef848abdd808509a7d82973af1197fff72ba4575e319c8"} Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.201409 4777 scope.go:117] "RemoveContainer" containerID="05e21734aec05ff774fe535fea9c31c3c6dafffa7b3c7ac51f401de3cd19fbba" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.201508 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.257202 4777 scope.go:117] "RemoveContainer" containerID="e64c61f2703e560ea489d9bd08de334147c4967a65951dfb9b2c81d4873c35dc" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.269746 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.277947 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.286203 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 17:20:07 crc kubenswrapper[4777]: E1124 17:20:07.286752 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerName="prometheus" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.286764 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerName="prometheus" Nov 24 17:20:07 crc kubenswrapper[4777]: E1124 17:20:07.286782 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerName="config-reloader" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.286788 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerName="config-reloader" Nov 24 17:20:07 crc kubenswrapper[4777]: E1124 17:20:07.286803 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerName="init-config-reloader" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.286809 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerName="init-config-reloader" Nov 24 17:20:07 crc kubenswrapper[4777]: E1124 17:20:07.286820 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerName="thanos-sidecar" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.286826 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerName="thanos-sidecar" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.287004 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerName="config-reloader" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.287017 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerName="prometheus" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.287039 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" containerName="thanos-sidecar" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.289178 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.297216 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.297232 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.297540 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.298241 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.298301 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.300084 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-mxt75" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.300113 4777 scope.go:117] "RemoveContainer" containerID="8bc55352e58db3657ba1188879f371842295e1d4d4f5233b6c23347da8f5c05e" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.300594 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.321142 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.343259 4777 scope.go:117] "RemoveContainer" containerID="efb7c66a1887ce7a1d81fe309029105ee9be5adcb1357020783c58d1c95679c7" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.405933 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.406204 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.406315 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/df409568-e0d5-49db-8da4-efbe45ec3a2c-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.406369 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqvxg\" (UniqueName: \"kubernetes.io/projected/df409568-e0d5-49db-8da4-efbe45ec3a2c-kube-api-access-jqvxg\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.406442 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-872360a4-14ac-4955-839d-37d8c1b22667\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-872360a4-14ac-4955-839d-37d8c1b22667\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.406510 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.406564 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.406701 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.406764 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-config\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.406782 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/df409568-e0d5-49db-8da4-efbe45ec3a2c-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.406863 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/df409568-e0d5-49db-8da4-efbe45ec3a2c-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.514655 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.514719 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.514762 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.514792 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-config\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.514812 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/df409568-e0d5-49db-8da4-efbe45ec3a2c-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.514853 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/df409568-e0d5-49db-8da4-efbe45ec3a2c-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.514906 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.514926 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.514990 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/df409568-e0d5-49db-8da4-efbe45ec3a2c-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.515014 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqvxg\" (UniqueName: \"kubernetes.io/projected/df409568-e0d5-49db-8da4-efbe45ec3a2c-kube-api-access-jqvxg\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.515050 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-872360a4-14ac-4955-839d-37d8c1b22667\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-872360a4-14ac-4955-839d-37d8c1b22667\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.520108 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/df409568-e0d5-49db-8da4-efbe45ec3a2c-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.521798 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/df409568-e0d5-49db-8da4-efbe45ec3a2c-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.523449 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.523812 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.529863 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.530483 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.531007 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-config\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.532186 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/df409568-e0d5-49db-8da4-efbe45ec3a2c-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.536911 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.536984 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-872360a4-14ac-4955-839d-37d8c1b22667\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-872360a4-14ac-4955-839d-37d8c1b22667\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b9e58ef1499554db3d4c243d9997bf278b9eaa38e6e18118db34549346e6e461/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.540473 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqvxg\" (UniqueName: \"kubernetes.io/projected/df409568-e0d5-49db-8da4-efbe45ec3a2c-kube-api-access-jqvxg\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.564638 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/df409568-e0d5-49db-8da4-efbe45ec3a2c-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.615210 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-872360a4-14ac-4955-839d-37d8c1b22667\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-872360a4-14ac-4955-839d-37d8c1b22667\") pod \"prometheus-metric-storage-0\" (UID: \"df409568-e0d5-49db-8da4-efbe45ec3a2c\") " pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:07 crc kubenswrapper[4777]: I1124 17:20:07.926464 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:08 crc kubenswrapper[4777]: I1124 17:20:08.421910 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 17:20:08 crc kubenswrapper[4777]: I1124 17:20:08.732631 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="f68bfbdc-e400-488d-9f32-ea90fd1160e3" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 17:20:09 crc kubenswrapper[4777]: I1124 17:20:09.227343 4777 generic.go:334] "Generic (PLEG): container finished" podID="3ec24d0c-556b-4a24-8f42-38fed08efe6b" containerID="62ffc2935295b2dbd24c5488e3bdb6d73839009626d5cf6cd3810e31fc0aed49" exitCode=0 Nov 24 17:20:09 crc kubenswrapper[4777]: I1124 17:20:09.227636 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xgm6s" event={"ID":"3ec24d0c-556b-4a24-8f42-38fed08efe6b","Type":"ContainerDied","Data":"62ffc2935295b2dbd24c5488e3bdb6d73839009626d5cf6cd3810e31fc0aed49"} Nov 24 17:20:09 crc kubenswrapper[4777]: I1124 17:20:09.232110 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"df409568-e0d5-49db-8da4-efbe45ec3a2c","Type":"ContainerStarted","Data":"04304be2f447d4626ab471f563fee3a203f3358f946b2ec89d654d63d7b934ab"} Nov 24 17:20:09 crc kubenswrapper[4777]: I1124 17:20:09.257801 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9ab961d-98f3-4a5e-9583-a168fa95f7b4" path="/var/lib/kubelet/pods/f9ab961d-98f3-4a5e-9583-a168fa95f7b4/volumes" Nov 24 17:20:10 crc kubenswrapper[4777]: I1124 17:20:10.702590 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:20:10 crc kubenswrapper[4777]: I1124 17:20:10.805517 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-q4l2k" podUID="73928e73-667c-4b69-aed9-72f32012fbdc" containerName="ovn-controller" probeResult="failure" output=< Nov 24 17:20:10 crc kubenswrapper[4777]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 17:20:10 crc kubenswrapper[4777]: > Nov 24 17:20:10 crc kubenswrapper[4777]: I1124 17:20:10.885471 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-combined-ca-bundle\") pod \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " Nov 24 17:20:10 crc kubenswrapper[4777]: I1124 17:20:10.885569 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75kwr\" (UniqueName: \"kubernetes.io/projected/3ec24d0c-556b-4a24-8f42-38fed08efe6b-kube-api-access-75kwr\") pod \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " Nov 24 17:20:10 crc kubenswrapper[4777]: I1124 17:20:10.885592 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-dispersionconf\") pod \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " Nov 24 17:20:10 crc kubenswrapper[4777]: I1124 17:20:10.885638 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ec24d0c-556b-4a24-8f42-38fed08efe6b-scripts\") pod \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " Nov 24 17:20:10 crc kubenswrapper[4777]: I1124 17:20:10.885662 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-swiftconf\") pod \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " Nov 24 17:20:10 crc kubenswrapper[4777]: I1124 17:20:10.885690 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ec24d0c-556b-4a24-8f42-38fed08efe6b-ring-data-devices\") pod \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " Nov 24 17:20:10 crc kubenswrapper[4777]: I1124 17:20:10.885756 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ec24d0c-556b-4a24-8f42-38fed08efe6b-etc-swift\") pod \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\" (UID: \"3ec24d0c-556b-4a24-8f42-38fed08efe6b\") " Nov 24 17:20:10 crc kubenswrapper[4777]: I1124 17:20:10.886831 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ec24d0c-556b-4a24-8f42-38fed08efe6b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3ec24d0c-556b-4a24-8f42-38fed08efe6b" (UID: "3ec24d0c-556b-4a24-8f42-38fed08efe6b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:20:10 crc kubenswrapper[4777]: I1124 17:20:10.888572 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ec24d0c-556b-4a24-8f42-38fed08efe6b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3ec24d0c-556b-4a24-8f42-38fed08efe6b" (UID: "3ec24d0c-556b-4a24-8f42-38fed08efe6b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:10 crc kubenswrapper[4777]: I1124 17:20:10.988716 4777 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ec24d0c-556b-4a24-8f42-38fed08efe6b-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:10 crc kubenswrapper[4777]: I1124 17:20:10.988752 4777 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ec24d0c-556b-4a24-8f42-38fed08efe6b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:11 crc kubenswrapper[4777]: I1124 17:20:11.047521 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ec24d0c-556b-4a24-8f42-38fed08efe6b-kube-api-access-75kwr" (OuterVolumeSpecName: "kube-api-access-75kwr") pod "3ec24d0c-556b-4a24-8f42-38fed08efe6b" (UID: "3ec24d0c-556b-4a24-8f42-38fed08efe6b"). InnerVolumeSpecName "kube-api-access-75kwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:11 crc kubenswrapper[4777]: I1124 17:20:11.049378 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ec24d0c-556b-4a24-8f42-38fed08efe6b-scripts" (OuterVolumeSpecName: "scripts") pod "3ec24d0c-556b-4a24-8f42-38fed08efe6b" (UID: "3ec24d0c-556b-4a24-8f42-38fed08efe6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:11 crc kubenswrapper[4777]: I1124 17:20:11.052639 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3ec24d0c-556b-4a24-8f42-38fed08efe6b" (UID: "3ec24d0c-556b-4a24-8f42-38fed08efe6b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:11 crc kubenswrapper[4777]: I1124 17:20:11.055060 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ec24d0c-556b-4a24-8f42-38fed08efe6b" (UID: "3ec24d0c-556b-4a24-8f42-38fed08efe6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:11 crc kubenswrapper[4777]: I1124 17:20:11.065051 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3ec24d0c-556b-4a24-8f42-38fed08efe6b" (UID: "3ec24d0c-556b-4a24-8f42-38fed08efe6b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:11 crc kubenswrapper[4777]: I1124 17:20:11.089823 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:11 crc kubenswrapper[4777]: I1124 17:20:11.089851 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75kwr\" (UniqueName: \"kubernetes.io/projected/3ec24d0c-556b-4a24-8f42-38fed08efe6b-kube-api-access-75kwr\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:11 crc kubenswrapper[4777]: I1124 17:20:11.089865 4777 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:11 crc kubenswrapper[4777]: I1124 17:20:11.089876 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ec24d0c-556b-4a24-8f42-38fed08efe6b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:11 crc kubenswrapper[4777]: I1124 17:20:11.089887 4777 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ec24d0c-556b-4a24-8f42-38fed08efe6b-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:11 crc kubenswrapper[4777]: I1124 17:20:11.278210 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xgm6s" event={"ID":"3ec24d0c-556b-4a24-8f42-38fed08efe6b","Type":"ContainerDied","Data":"6a410e1deeef1f929bf9dbcfec49f6ad8859a7ced740439460ddfb36e67b0784"} Nov 24 17:20:11 crc kubenswrapper[4777]: I1124 17:20:11.278797 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a410e1deeef1f929bf9dbcfec49f6ad8859a7ced740439460ddfb36e67b0784" Nov 24 17:20:11 crc kubenswrapper[4777]: I1124 17:20:11.278942 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xgm6s" Nov 24 17:20:12 crc kubenswrapper[4777]: I1124 17:20:12.070616 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 24 17:20:12 crc kubenswrapper[4777]: I1124 17:20:12.286396 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"df409568-e0d5-49db-8da4-efbe45ec3a2c","Type":"ContainerStarted","Data":"968f4b10be60a867d4f2f3b83e3b713f1adfa2cbc08fe61ed4c01760afcf4624"} Nov 24 17:20:15 crc kubenswrapper[4777]: I1124 17:20:15.709576 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-q4l2k" podUID="73928e73-667c-4b69-aed9-72f32012fbdc" containerName="ovn-controller" probeResult="failure" output=< Nov 24 17:20:15 crc kubenswrapper[4777]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 17:20:15 crc kubenswrapper[4777]: > Nov 24 17:20:15 crc kubenswrapper[4777]: I1124 17:20:15.792203 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:20:15 crc kubenswrapper[4777]: I1124 17:20:15.795183 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-dkbds" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.023268 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-q4l2k-config-mzkfx"] Nov 24 17:20:16 crc kubenswrapper[4777]: E1124 17:20:16.023702 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ec24d0c-556b-4a24-8f42-38fed08efe6b" containerName="swift-ring-rebalance" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.023718 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ec24d0c-556b-4a24-8f42-38fed08efe6b" containerName="swift-ring-rebalance" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.023942 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ec24d0c-556b-4a24-8f42-38fed08efe6b" containerName="swift-ring-rebalance" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.024724 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.030851 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.047655 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-q4l2k-config-mzkfx"] Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.185841 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-run\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.185893 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-run-ovn\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.186006 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7f6109d2-ed42-48f2-8998-f14ab220416f-additional-scripts\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.186062 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqrps\" (UniqueName: \"kubernetes.io/projected/7f6109d2-ed42-48f2-8998-f14ab220416f-kube-api-access-sqrps\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.186102 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f6109d2-ed42-48f2-8998-f14ab220416f-scripts\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.186133 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-log-ovn\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.287696 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-run\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.288295 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-run-ovn\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.288392 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7f6109d2-ed42-48f2-8998-f14ab220416f-additional-scripts\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.288486 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqrps\" (UniqueName: \"kubernetes.io/projected/7f6109d2-ed42-48f2-8998-f14ab220416f-kube-api-access-sqrps\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.288575 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f6109d2-ed42-48f2-8998-f14ab220416f-scripts\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.288647 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-log-ovn\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.288776 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-log-ovn\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.288173 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-run\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.288921 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-run-ovn\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.289544 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7f6109d2-ed42-48f2-8998-f14ab220416f-additional-scripts\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.291378 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f6109d2-ed42-48f2-8998-f14ab220416f-scripts\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.325253 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqrps\" (UniqueName: \"kubernetes.io/projected/7f6109d2-ed42-48f2-8998-f14ab220416f-kube-api-access-sqrps\") pod \"ovn-controller-q4l2k-config-mzkfx\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:16 crc kubenswrapper[4777]: I1124 17:20:16.348940 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:18 crc kubenswrapper[4777]: I1124 17:20:18.720831 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-ingester-0" Nov 24 17:20:19 crc kubenswrapper[4777]: I1124 17:20:19.353540 4777 generic.go:334] "Generic (PLEG): container finished" podID="df409568-e0d5-49db-8da4-efbe45ec3a2c" containerID="968f4b10be60a867d4f2f3b83e3b713f1adfa2cbc08fe61ed4c01760afcf4624" exitCode=0 Nov 24 17:20:19 crc kubenswrapper[4777]: I1124 17:20:19.353638 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"df409568-e0d5-49db-8da4-efbe45ec3a2c","Type":"ContainerDied","Data":"968f4b10be60a867d4f2f3b83e3b713f1adfa2cbc08fe61ed4c01760afcf4624"} Nov 24 17:20:20 crc kubenswrapper[4777]: I1124 17:20:20.735935 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-q4l2k" podUID="73928e73-667c-4b69-aed9-72f32012fbdc" containerName="ovn-controller" probeResult="failure" output=< Nov 24 17:20:20 crc kubenswrapper[4777]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 17:20:20 crc kubenswrapper[4777]: > Nov 24 17:20:20 crc kubenswrapper[4777]: I1124 17:20:20.890445 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-q4l2k-config-mzkfx"] Nov 24 17:20:20 crc kubenswrapper[4777]: W1124 17:20:20.900000 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f6109d2_ed42_48f2_8998_f14ab220416f.slice/crio-444b4970f3a3a0989db90b22c605d40b7297ba929a55b73028c1d8239c0c43fc WatchSource:0}: Error finding container 444b4970f3a3a0989db90b22c605d40b7297ba929a55b73028c1d8239c0c43fc: Status 404 returned error can't find the container with id 444b4970f3a3a0989db90b22c605d40b7297ba929a55b73028c1d8239c0c43fc Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.343382 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.384661 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-q4l2k-config-mzkfx" event={"ID":"7f6109d2-ed42-48f2-8998-f14ab220416f","Type":"ContainerStarted","Data":"1c6e63bcd736b4e0d34d704190b35d5a66b8cc042af83e7c9a57888c102700a2"} Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.384717 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-q4l2k-config-mzkfx" event={"ID":"7f6109d2-ed42-48f2-8998-f14ab220416f","Type":"ContainerStarted","Data":"444b4970f3a3a0989db90b22c605d40b7297ba929a55b73028c1d8239c0c43fc"} Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.386812 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"df409568-e0d5-49db-8da4-efbe45ec3a2c","Type":"ContainerStarted","Data":"6d9d2a3cb4e9023fdc44cf030976ea6124a59f70fdb484fec9b3d4fbd7166ea7"} Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.397681 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-swrgj" event={"ID":"7b60b793-78a5-406c-9ed6-e582dfba568b","Type":"ContainerStarted","Data":"d57f5c5e958e3c77e35416bfd8f87db04e97a311ca0c6689b7decfa910bf7618"} Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.445462 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-q4l2k-config-mzkfx" podStartSLOduration=5.445441458 podStartE2EDuration="5.445441458s" podCreationTimestamp="2025-11-24 17:20:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:20:21.403184783 +0000 UTC m=+1179.562379852" watchObservedRunningTime="2025-11-24 17:20:21.445441458 +0000 UTC m=+1179.604636507" Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.452797 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-swrgj" podStartSLOduration=2.475949332 podStartE2EDuration="16.452779249s" podCreationTimestamp="2025-11-24 17:20:05 +0000 UTC" firstStartedPulling="2025-11-24 17:20:06.45336886 +0000 UTC m=+1164.612563909" lastFinishedPulling="2025-11-24 17:20:20.430198737 +0000 UTC m=+1178.589393826" observedRunningTime="2025-11-24 17:20:21.441623819 +0000 UTC m=+1179.600818878" watchObservedRunningTime="2025-11-24 17:20:21.452779249 +0000 UTC m=+1179.611974298" Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.610130 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.802156 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.841602 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9c412d58-6198-4254-932d-5eb90420f8ba-etc-swift\") pod \"swift-storage-0\" (UID: \"9c412d58-6198-4254-932d-5eb90420f8ba\") " pod="openstack/swift-storage-0" Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.843734 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-9666p"] Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.853321 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9666p" Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.859422 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-c352-account-create-r4cp5"] Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.868319 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c352-account-create-r4cp5" Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.891561 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.918050 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-9666p"] Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.937152 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c352-account-create-r4cp5"] Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.953364 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.987027 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-create-86vqg"] Nov 24 17:20:21 crc kubenswrapper[4777]: I1124 17:20:21.988496 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-86vqg" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.012069 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab4c523b-a2f8-4c22-bbf8-c4ce606bd932-operator-scripts\") pod \"barbican-c352-account-create-r4cp5\" (UID: \"ab4c523b-a2f8-4c22-bbf8-c4ce606bd932\") " pod="openstack/barbican-c352-account-create-r4cp5" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.012109 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xk8f\" (UniqueName: \"kubernetes.io/projected/ab4c523b-a2f8-4c22-bbf8-c4ce606bd932-kube-api-access-5xk8f\") pod \"barbican-c352-account-create-r4cp5\" (UID: \"ab4c523b-a2f8-4c22-bbf8-c4ce606bd932\") " pod="openstack/barbican-c352-account-create-r4cp5" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.012169 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d7c6a9d-cea0-4fab-abc8-53e1926587cc-operator-scripts\") pod \"cinder-db-create-9666p\" (UID: \"8d7c6a9d-cea0-4fab-abc8-53e1926587cc\") " pod="openstack/cinder-db-create-9666p" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.012247 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld6bb\" (UniqueName: \"kubernetes.io/projected/8d7c6a9d-cea0-4fab-abc8-53e1926587cc-kube-api-access-ld6bb\") pod \"cinder-db-create-9666p\" (UID: \"8d7c6a9d-cea0-4fab-abc8-53e1926587cc\") " pod="openstack/cinder-db-create-9666p" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.012833 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-5d5f-account-create-vxk8c"] Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.013991 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5d5f-account-create-vxk8c" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.023662 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.028039 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-create-86vqg"] Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.045648 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5d5f-account-create-vxk8c"] Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.078649 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-q9djb"] Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.080332 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-q9djb" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.114061 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d7c6a9d-cea0-4fab-abc8-53e1926587cc-operator-scripts\") pod \"cinder-db-create-9666p\" (UID: \"8d7c6a9d-cea0-4fab-abc8-53e1926587cc\") " pod="openstack/cinder-db-create-9666p" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.114313 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf61c287-5b51-4b32-a50e-54b93835632d-operator-scripts\") pod \"cloudkitty-db-create-86vqg\" (UID: \"bf61c287-5b51-4b32-a50e-54b93835632d\") " pod="openstack/cloudkitty-db-create-86vqg" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.114387 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6389225-005e-4032-a3bb-a5700e55d08e-operator-scripts\") pod \"cinder-5d5f-account-create-vxk8c\" (UID: \"c6389225-005e-4032-a3bb-a5700e55d08e\") " pod="openstack/cinder-5d5f-account-create-vxk8c" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.114503 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlsxz\" (UniqueName: \"kubernetes.io/projected/c6389225-005e-4032-a3bb-a5700e55d08e-kube-api-access-zlsxz\") pod \"cinder-5d5f-account-create-vxk8c\" (UID: \"c6389225-005e-4032-a3bb-a5700e55d08e\") " pod="openstack/cinder-5d5f-account-create-vxk8c" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.114586 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld6bb\" (UniqueName: \"kubernetes.io/projected/8d7c6a9d-cea0-4fab-abc8-53e1926587cc-kube-api-access-ld6bb\") pod \"cinder-db-create-9666p\" (UID: \"8d7c6a9d-cea0-4fab-abc8-53e1926587cc\") " pod="openstack/cinder-db-create-9666p" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.114695 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khksx\" (UniqueName: \"kubernetes.io/projected/bf61c287-5b51-4b32-a50e-54b93835632d-kube-api-access-khksx\") pod \"cloudkitty-db-create-86vqg\" (UID: \"bf61c287-5b51-4b32-a50e-54b93835632d\") " pod="openstack/cloudkitty-db-create-86vqg" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.114798 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab4c523b-a2f8-4c22-bbf8-c4ce606bd932-operator-scripts\") pod \"barbican-c352-account-create-r4cp5\" (UID: \"ab4c523b-a2f8-4c22-bbf8-c4ce606bd932\") " pod="openstack/barbican-c352-account-create-r4cp5" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.114862 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xk8f\" (UniqueName: \"kubernetes.io/projected/ab4c523b-a2f8-4c22-bbf8-c4ce606bd932-kube-api-access-5xk8f\") pod \"barbican-c352-account-create-r4cp5\" (UID: \"ab4c523b-a2f8-4c22-bbf8-c4ce606bd932\") " pod="openstack/barbican-c352-account-create-r4cp5" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.115828 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d7c6a9d-cea0-4fab-abc8-53e1926587cc-operator-scripts\") pod \"cinder-db-create-9666p\" (UID: \"8d7c6a9d-cea0-4fab-abc8-53e1926587cc\") " pod="openstack/cinder-db-create-9666p" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.117288 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab4c523b-a2f8-4c22-bbf8-c4ce606bd932-operator-scripts\") pod \"barbican-c352-account-create-r4cp5\" (UID: \"ab4c523b-a2f8-4c22-bbf8-c4ce606bd932\") " pod="openstack/barbican-c352-account-create-r4cp5" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.122897 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-q9djb"] Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.216841 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf61c287-5b51-4b32-a50e-54b93835632d-operator-scripts\") pod \"cloudkitty-db-create-86vqg\" (UID: \"bf61c287-5b51-4b32-a50e-54b93835632d\") " pod="openstack/cloudkitty-db-create-86vqg" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.217228 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6389225-005e-4032-a3bb-a5700e55d08e-operator-scripts\") pod \"cinder-5d5f-account-create-vxk8c\" (UID: \"c6389225-005e-4032-a3bb-a5700e55d08e\") " pod="openstack/cinder-5d5f-account-create-vxk8c" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.217357 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e082ddb3-c5ea-4e65-b002-abefd538a16b-operator-scripts\") pod \"barbican-db-create-q9djb\" (UID: \"e082ddb3-c5ea-4e65-b002-abefd538a16b\") " pod="openstack/barbican-db-create-q9djb" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.217494 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlsxz\" (UniqueName: \"kubernetes.io/projected/c6389225-005e-4032-a3bb-a5700e55d08e-kube-api-access-zlsxz\") pod \"cinder-5d5f-account-create-vxk8c\" (UID: \"c6389225-005e-4032-a3bb-a5700e55d08e\") " pod="openstack/cinder-5d5f-account-create-vxk8c" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.217622 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c268k\" (UniqueName: \"kubernetes.io/projected/e082ddb3-c5ea-4e65-b002-abefd538a16b-kube-api-access-c268k\") pod \"barbican-db-create-q9djb\" (UID: \"e082ddb3-c5ea-4e65-b002-abefd538a16b\") " pod="openstack/barbican-db-create-q9djb" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.217794 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khksx\" (UniqueName: \"kubernetes.io/projected/bf61c287-5b51-4b32-a50e-54b93835632d-kube-api-access-khksx\") pod \"cloudkitty-db-create-86vqg\" (UID: \"bf61c287-5b51-4b32-a50e-54b93835632d\") " pod="openstack/cloudkitty-db-create-86vqg" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.219236 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf61c287-5b51-4b32-a50e-54b93835632d-operator-scripts\") pod \"cloudkitty-db-create-86vqg\" (UID: \"bf61c287-5b51-4b32-a50e-54b93835632d\") " pod="openstack/cloudkitty-db-create-86vqg" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.219845 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6389225-005e-4032-a3bb-a5700e55d08e-operator-scripts\") pod \"cinder-5d5f-account-create-vxk8c\" (UID: \"c6389225-005e-4032-a3bb-a5700e55d08e\") " pod="openstack/cinder-5d5f-account-create-vxk8c" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.237459 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-stwcx"] Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.238803 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-stwcx" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.254121 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.254262 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xk8f\" (UniqueName: \"kubernetes.io/projected/ab4c523b-a2f8-4c22-bbf8-c4ce606bd932-kube-api-access-5xk8f\") pod \"barbican-c352-account-create-r4cp5\" (UID: \"ab4c523b-a2f8-4c22-bbf8-c4ce606bd932\") " pod="openstack/barbican-c352-account-create-r4cp5" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.254317 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.254562 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vzrxj" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.254742 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.255390 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-stwcx"] Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.257749 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld6bb\" (UniqueName: \"kubernetes.io/projected/8d7c6a9d-cea0-4fab-abc8-53e1926587cc-kube-api-access-ld6bb\") pod \"cinder-db-create-9666p\" (UID: \"8d7c6a9d-cea0-4fab-abc8-53e1926587cc\") " pod="openstack/cinder-db-create-9666p" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.261594 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlsxz\" (UniqueName: \"kubernetes.io/projected/c6389225-005e-4032-a3bb-a5700e55d08e-kube-api-access-zlsxz\") pod \"cinder-5d5f-account-create-vxk8c\" (UID: \"c6389225-005e-4032-a3bb-a5700e55d08e\") " pod="openstack/cinder-5d5f-account-create-vxk8c" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.272235 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khksx\" (UniqueName: \"kubernetes.io/projected/bf61c287-5b51-4b32-a50e-54b93835632d-kube-api-access-khksx\") pod \"cloudkitty-db-create-86vqg\" (UID: \"bf61c287-5b51-4b32-a50e-54b93835632d\") " pod="openstack/cloudkitty-db-create-86vqg" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.324201 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc393164-6269-4238-8eab-294157bd3e0a-config-data\") pod \"keystone-db-sync-stwcx\" (UID: \"cc393164-6269-4238-8eab-294157bd3e0a\") " pod="openstack/keystone-db-sync-stwcx" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.324626 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc393164-6269-4238-8eab-294157bd3e0a-combined-ca-bundle\") pod \"keystone-db-sync-stwcx\" (UID: \"cc393164-6269-4238-8eab-294157bd3e0a\") " pod="openstack/keystone-db-sync-stwcx" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.324727 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e082ddb3-c5ea-4e65-b002-abefd538a16b-operator-scripts\") pod \"barbican-db-create-q9djb\" (UID: \"e082ddb3-c5ea-4e65-b002-abefd538a16b\") " pod="openstack/barbican-db-create-q9djb" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.324764 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-492bz\" (UniqueName: \"kubernetes.io/projected/cc393164-6269-4238-8eab-294157bd3e0a-kube-api-access-492bz\") pod \"keystone-db-sync-stwcx\" (UID: \"cc393164-6269-4238-8eab-294157bd3e0a\") " pod="openstack/keystone-db-sync-stwcx" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.324781 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c268k\" (UniqueName: \"kubernetes.io/projected/e082ddb3-c5ea-4e65-b002-abefd538a16b-kube-api-access-c268k\") pod \"barbican-db-create-q9djb\" (UID: \"e082ddb3-c5ea-4e65-b002-abefd538a16b\") " pod="openstack/barbican-db-create-q9djb" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.325951 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e082ddb3-c5ea-4e65-b002-abefd538a16b-operator-scripts\") pod \"barbican-db-create-q9djb\" (UID: \"e082ddb3-c5ea-4e65-b002-abefd538a16b\") " pod="openstack/barbican-db-create-q9djb" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.351547 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c268k\" (UniqueName: \"kubernetes.io/projected/e082ddb3-c5ea-4e65-b002-abefd538a16b-kube-api-access-c268k\") pod \"barbican-db-create-q9djb\" (UID: \"e082ddb3-c5ea-4e65-b002-abefd538a16b\") " pod="openstack/barbican-db-create-q9djb" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.358221 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-fzsq6"] Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.359528 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fzsq6" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.374222 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-590a-account-create-kxt2z"] Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.375458 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-590a-account-create-kxt2z" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.378550 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.394422 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-fzsq6"] Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.417185 4777 generic.go:334] "Generic (PLEG): container finished" podID="7f6109d2-ed42-48f2-8998-f14ab220416f" containerID="1c6e63bcd736b4e0d34d704190b35d5a66b8cc042af83e7c9a57888c102700a2" exitCode=0 Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.417777 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-q4l2k-config-mzkfx" event={"ID":"7f6109d2-ed42-48f2-8998-f14ab220416f","Type":"ContainerDied","Data":"1c6e63bcd736b4e0d34d704190b35d5a66b8cc042af83e7c9a57888c102700a2"} Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.427237 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-590a-account-create-kxt2z"] Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.430276 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-492bz\" (UniqueName: \"kubernetes.io/projected/cc393164-6269-4238-8eab-294157bd3e0a-kube-api-access-492bz\") pod \"keystone-db-sync-stwcx\" (UID: \"cc393164-6269-4238-8eab-294157bd3e0a\") " pod="openstack/keystone-db-sync-stwcx" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.430344 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc393164-6269-4238-8eab-294157bd3e0a-config-data\") pod \"keystone-db-sync-stwcx\" (UID: \"cc393164-6269-4238-8eab-294157bd3e0a\") " pod="openstack/keystone-db-sync-stwcx" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.430416 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc393164-6269-4238-8eab-294157bd3e0a-combined-ca-bundle\") pod \"keystone-db-sync-stwcx\" (UID: \"cc393164-6269-4238-8eab-294157bd3e0a\") " pod="openstack/keystone-db-sync-stwcx" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.433851 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc393164-6269-4238-8eab-294157bd3e0a-config-data\") pod \"keystone-db-sync-stwcx\" (UID: \"cc393164-6269-4238-8eab-294157bd3e0a\") " pod="openstack/keystone-db-sync-stwcx" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.441872 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc393164-6269-4238-8eab-294157bd3e0a-combined-ca-bundle\") pod \"keystone-db-sync-stwcx\" (UID: \"cc393164-6269-4238-8eab-294157bd3e0a\") " pod="openstack/keystone-db-sync-stwcx" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.472489 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-492bz\" (UniqueName: \"kubernetes.io/projected/cc393164-6269-4238-8eab-294157bd3e0a-kube-api-access-492bz\") pod \"keystone-db-sync-stwcx\" (UID: \"cc393164-6269-4238-8eab-294157bd3e0a\") " pod="openstack/keystone-db-sync-stwcx" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.472927 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-86vqg" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.476433 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5d5f-account-create-vxk8c" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.512287 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-q9djb" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.532640 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81ec0e3a-1906-4013-a5cd-c639933a26a5-operator-scripts\") pod \"neutron-590a-account-create-kxt2z\" (UID: \"81ec0e3a-1906-4013-a5cd-c639933a26a5\") " pod="openstack/neutron-590a-account-create-kxt2z" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.532766 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6jmx\" (UniqueName: \"kubernetes.io/projected/81ec0e3a-1906-4013-a5cd-c639933a26a5-kube-api-access-k6jmx\") pod \"neutron-590a-account-create-kxt2z\" (UID: \"81ec0e3a-1906-4013-a5cd-c639933a26a5\") " pod="openstack/neutron-590a-account-create-kxt2z" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.532805 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c41309a-1061-4cf2-bf89-f170ee446905-operator-scripts\") pod \"neutron-db-create-fzsq6\" (UID: \"7c41309a-1061-4cf2-bf89-f170ee446905\") " pod="openstack/neutron-db-create-fzsq6" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.532857 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxgbk\" (UniqueName: \"kubernetes.io/projected/7c41309a-1061-4cf2-bf89-f170ee446905-kube-api-access-bxgbk\") pod \"neutron-db-create-fzsq6\" (UID: \"7c41309a-1061-4cf2-bf89-f170ee446905\") " pod="openstack/neutron-db-create-fzsq6" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.537807 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9666p" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.552609 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c352-account-create-r4cp5" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.554872 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-19af-account-create-9fl7d"] Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.556931 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-19af-account-create-9fl7d" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.567700 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-db-secret" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.582691 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-stwcx" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.594961 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-19af-account-create-9fl7d"] Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.635195 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6jmx\" (UniqueName: \"kubernetes.io/projected/81ec0e3a-1906-4013-a5cd-c639933a26a5-kube-api-access-k6jmx\") pod \"neutron-590a-account-create-kxt2z\" (UID: \"81ec0e3a-1906-4013-a5cd-c639933a26a5\") " pod="openstack/neutron-590a-account-create-kxt2z" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.635262 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c41309a-1061-4cf2-bf89-f170ee446905-operator-scripts\") pod \"neutron-db-create-fzsq6\" (UID: \"7c41309a-1061-4cf2-bf89-f170ee446905\") " pod="openstack/neutron-db-create-fzsq6" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.635314 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9721c23b-c04c-4dc4-94ab-24455982fbae-operator-scripts\") pod \"cloudkitty-19af-account-create-9fl7d\" (UID: \"9721c23b-c04c-4dc4-94ab-24455982fbae\") " pod="openstack/cloudkitty-19af-account-create-9fl7d" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.635362 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxgbk\" (UniqueName: \"kubernetes.io/projected/7c41309a-1061-4cf2-bf89-f170ee446905-kube-api-access-bxgbk\") pod \"neutron-db-create-fzsq6\" (UID: \"7c41309a-1061-4cf2-bf89-f170ee446905\") " pod="openstack/neutron-db-create-fzsq6" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.635409 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f527\" (UniqueName: \"kubernetes.io/projected/9721c23b-c04c-4dc4-94ab-24455982fbae-kube-api-access-2f527\") pod \"cloudkitty-19af-account-create-9fl7d\" (UID: \"9721c23b-c04c-4dc4-94ab-24455982fbae\") " pod="openstack/cloudkitty-19af-account-create-9fl7d" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.635487 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81ec0e3a-1906-4013-a5cd-c639933a26a5-operator-scripts\") pod \"neutron-590a-account-create-kxt2z\" (UID: \"81ec0e3a-1906-4013-a5cd-c639933a26a5\") " pod="openstack/neutron-590a-account-create-kxt2z" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.636590 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81ec0e3a-1906-4013-a5cd-c639933a26a5-operator-scripts\") pod \"neutron-590a-account-create-kxt2z\" (UID: \"81ec0e3a-1906-4013-a5cd-c639933a26a5\") " pod="openstack/neutron-590a-account-create-kxt2z" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.637132 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c41309a-1061-4cf2-bf89-f170ee446905-operator-scripts\") pod \"neutron-db-create-fzsq6\" (UID: \"7c41309a-1061-4cf2-bf89-f170ee446905\") " pod="openstack/neutron-db-create-fzsq6" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.668426 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6jmx\" (UniqueName: \"kubernetes.io/projected/81ec0e3a-1906-4013-a5cd-c639933a26a5-kube-api-access-k6jmx\") pod \"neutron-590a-account-create-kxt2z\" (UID: \"81ec0e3a-1906-4013-a5cd-c639933a26a5\") " pod="openstack/neutron-590a-account-create-kxt2z" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.709698 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxgbk\" (UniqueName: \"kubernetes.io/projected/7c41309a-1061-4cf2-bf89-f170ee446905-kube-api-access-bxgbk\") pod \"neutron-db-create-fzsq6\" (UID: \"7c41309a-1061-4cf2-bf89-f170ee446905\") " pod="openstack/neutron-db-create-fzsq6" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.743931 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9721c23b-c04c-4dc4-94ab-24455982fbae-operator-scripts\") pod \"cloudkitty-19af-account-create-9fl7d\" (UID: \"9721c23b-c04c-4dc4-94ab-24455982fbae\") " pod="openstack/cloudkitty-19af-account-create-9fl7d" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.744082 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f527\" (UniqueName: \"kubernetes.io/projected/9721c23b-c04c-4dc4-94ab-24455982fbae-kube-api-access-2f527\") pod \"cloudkitty-19af-account-create-9fl7d\" (UID: \"9721c23b-c04c-4dc4-94ab-24455982fbae\") " pod="openstack/cloudkitty-19af-account-create-9fl7d" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.747415 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9721c23b-c04c-4dc4-94ab-24455982fbae-operator-scripts\") pod \"cloudkitty-19af-account-create-9fl7d\" (UID: \"9721c23b-c04c-4dc4-94ab-24455982fbae\") " pod="openstack/cloudkitty-19af-account-create-9fl7d" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.767681 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f527\" (UniqueName: \"kubernetes.io/projected/9721c23b-c04c-4dc4-94ab-24455982fbae-kube-api-access-2f527\") pod \"cloudkitty-19af-account-create-9fl7d\" (UID: \"9721c23b-c04c-4dc4-94ab-24455982fbae\") " pod="openstack/cloudkitty-19af-account-create-9fl7d" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.783648 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-590a-account-create-kxt2z" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.787879 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-19af-account-create-9fl7d" Nov 24 17:20:22 crc kubenswrapper[4777]: I1124 17:20:22.985308 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fzsq6" Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.016096 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.126153 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5d5f-account-create-vxk8c"] Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.381065 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-q9djb"] Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.393753 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-create-86vqg"] Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.418492 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c352-account-create-r4cp5"] Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.449252 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-q9djb" event={"ID":"e082ddb3-c5ea-4e65-b002-abefd538a16b","Type":"ContainerStarted","Data":"3d2d4871f4112eb69eb0ebe7363b0be07a414a3dca8e02a640367921abaebefe"} Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.468658 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"df409568-e0d5-49db-8da4-efbe45ec3a2c","Type":"ContainerStarted","Data":"46d4149191c7300a4421c15bf32686026b66ee7d287a5b169d8b657ef7da713a"} Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.478016 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5d5f-account-create-vxk8c" event={"ID":"c6389225-005e-4032-a3bb-a5700e55d08e","Type":"ContainerStarted","Data":"1059575ea70673f1b2910ff87354c9f1af24c352a565c6f2732c2b173976a28e"} Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.485888 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c352-account-create-r4cp5" event={"ID":"ab4c523b-a2f8-4c22-bbf8-c4ce606bd932","Type":"ContainerStarted","Data":"a2c4a94985c2354ea916dd320969d00bacec5083490e57c8c5f182932fd8947e"} Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.499811 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-86vqg" event={"ID":"bf61c287-5b51-4b32-a50e-54b93835632d","Type":"ContainerStarted","Data":"b60ed3b0c047c0d88dc42bc9b7a61e29fd33ca1e50e118645ca12b7d72a5befc"} Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.516456 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"bff652fc2310ef8e38926a2dfb8e08ea6fa9d358cf8eb9dc1b48fb9867d95ccb"} Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.521824 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-9666p"] Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.528412 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-590a-account-create-kxt2z"] Nov 24 17:20:23 crc kubenswrapper[4777]: W1124 17:20:23.592519 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81ec0e3a_1906_4013_a5cd_c639933a26a5.slice/crio-5cfa0c1f99d8f58d88da3e21538736b3254b18004ca9ba63fbb2098d4ce8b351 WatchSource:0}: Error finding container 5cfa0c1f99d8f58d88da3e21538736b3254b18004ca9ba63fbb2098d4ce8b351: Status 404 returned error can't find the container with id 5cfa0c1f99d8f58d88da3e21538736b3254b18004ca9ba63fbb2098d4ce8b351 Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.680388 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-stwcx"] Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.721858 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-19af-account-create-9fl7d"] Nov 24 17:20:23 crc kubenswrapper[4777]: W1124 17:20:23.744335 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9721c23b_c04c_4dc4_94ab_24455982fbae.slice/crio-eb2656291667de58817ba71399a143d59fef18221423408e7e52b6d285b62187 WatchSource:0}: Error finding container eb2656291667de58817ba71399a143d59fef18221423408e7e52b6d285b62187: Status 404 returned error can't find the container with id eb2656291667de58817ba71399a143d59fef18221423408e7e52b6d285b62187 Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.766305 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-fzsq6"] Nov 24 17:20:23 crc kubenswrapper[4777]: I1124 17:20:23.979403 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.091147 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7f6109d2-ed42-48f2-8998-f14ab220416f-additional-scripts\") pod \"7f6109d2-ed42-48f2-8998-f14ab220416f\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.091259 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqrps\" (UniqueName: \"kubernetes.io/projected/7f6109d2-ed42-48f2-8998-f14ab220416f-kube-api-access-sqrps\") pod \"7f6109d2-ed42-48f2-8998-f14ab220416f\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.091303 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-run-ovn\") pod \"7f6109d2-ed42-48f2-8998-f14ab220416f\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.091363 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-run\") pod \"7f6109d2-ed42-48f2-8998-f14ab220416f\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.091429 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-log-ovn\") pod \"7f6109d2-ed42-48f2-8998-f14ab220416f\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.091455 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f6109d2-ed42-48f2-8998-f14ab220416f-scripts\") pod \"7f6109d2-ed42-48f2-8998-f14ab220416f\" (UID: \"7f6109d2-ed42-48f2-8998-f14ab220416f\") " Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.091916 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f6109d2-ed42-48f2-8998-f14ab220416f-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "7f6109d2-ed42-48f2-8998-f14ab220416f" (UID: "7f6109d2-ed42-48f2-8998-f14ab220416f"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.092172 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "7f6109d2-ed42-48f2-8998-f14ab220416f" (UID: "7f6109d2-ed42-48f2-8998-f14ab220416f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.092438 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-run" (OuterVolumeSpecName: "var-run") pod "7f6109d2-ed42-48f2-8998-f14ab220416f" (UID: "7f6109d2-ed42-48f2-8998-f14ab220416f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.092463 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "7f6109d2-ed42-48f2-8998-f14ab220416f" (UID: "7f6109d2-ed42-48f2-8998-f14ab220416f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.095015 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f6109d2-ed42-48f2-8998-f14ab220416f-scripts" (OuterVolumeSpecName: "scripts") pod "7f6109d2-ed42-48f2-8998-f14ab220416f" (UID: "7f6109d2-ed42-48f2-8998-f14ab220416f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.099607 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f6109d2-ed42-48f2-8998-f14ab220416f-kube-api-access-sqrps" (OuterVolumeSpecName: "kube-api-access-sqrps") pod "7f6109d2-ed42-48f2-8998-f14ab220416f" (UID: "7f6109d2-ed42-48f2-8998-f14ab220416f"). InnerVolumeSpecName "kube-api-access-sqrps". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.193293 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqrps\" (UniqueName: \"kubernetes.io/projected/7f6109d2-ed42-48f2-8998-f14ab220416f-kube-api-access-sqrps\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.193323 4777 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.193332 4777 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.193343 4777 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7f6109d2-ed42-48f2-8998-f14ab220416f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.193351 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f6109d2-ed42-48f2-8998-f14ab220416f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.193375 4777 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7f6109d2-ed42-48f2-8998-f14ab220416f-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.530655 4777 generic.go:334] "Generic (PLEG): container finished" podID="9721c23b-c04c-4dc4-94ab-24455982fbae" containerID="14ed858c11c45e2627d4e0b296f218ce37b196de7b5e5e448f4796979f0aea26" exitCode=0 Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.530740 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-19af-account-create-9fl7d" event={"ID":"9721c23b-c04c-4dc4-94ab-24455982fbae","Type":"ContainerDied","Data":"14ed858c11c45e2627d4e0b296f218ce37b196de7b5e5e448f4796979f0aea26"} Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.530769 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-19af-account-create-9fl7d" event={"ID":"9721c23b-c04c-4dc4-94ab-24455982fbae","Type":"ContainerStarted","Data":"eb2656291667de58817ba71399a143d59fef18221423408e7e52b6d285b62187"} Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.541704 4777 generic.go:334] "Generic (PLEG): container finished" podID="ab4c523b-a2f8-4c22-bbf8-c4ce606bd932" containerID="1f61570c43f2b5598bdc200b72f8b7fa54e80489f1203a96ca56ea88f7b25495" exitCode=0 Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.541809 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c352-account-create-r4cp5" event={"ID":"ab4c523b-a2f8-4c22-bbf8-c4ce606bd932","Type":"ContainerDied","Data":"1f61570c43f2b5598bdc200b72f8b7fa54e80489f1203a96ca56ea88f7b25495"} Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.543657 4777 generic.go:334] "Generic (PLEG): container finished" podID="bf61c287-5b51-4b32-a50e-54b93835632d" containerID="9d7cc59fd3983e9eeb50f7267b2de2bf999330eb8d1ba808e190de2c57944d28" exitCode=0 Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.543716 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-86vqg" event={"ID":"bf61c287-5b51-4b32-a50e-54b93835632d","Type":"ContainerDied","Data":"9d7cc59fd3983e9eeb50f7267b2de2bf999330eb8d1ba808e190de2c57944d28"} Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.557325 4777 generic.go:334] "Generic (PLEG): container finished" podID="8d7c6a9d-cea0-4fab-abc8-53e1926587cc" containerID="0bab90472c03fb88ec36c73d3b0f7c4207b59fca59be37a1911ceaf8e7e1a137" exitCode=0 Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.557406 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9666p" event={"ID":"8d7c6a9d-cea0-4fab-abc8-53e1926587cc","Type":"ContainerDied","Data":"0bab90472c03fb88ec36c73d3b0f7c4207b59fca59be37a1911ceaf8e7e1a137"} Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.557429 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9666p" event={"ID":"8d7c6a9d-cea0-4fab-abc8-53e1926587cc","Type":"ContainerStarted","Data":"9f09c3980739419e5631ad4050738c6180bb80a1ae4476e1a3e5d1879bd7274c"} Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.565517 4777 generic.go:334] "Generic (PLEG): container finished" podID="81ec0e3a-1906-4013-a5cd-c639933a26a5" containerID="1941bfbad606f73fef69beaa7014dccf63aa6793a0cc26746ac46ba0082e8e85" exitCode=0 Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.565581 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-590a-account-create-kxt2z" event={"ID":"81ec0e3a-1906-4013-a5cd-c639933a26a5","Type":"ContainerDied","Data":"1941bfbad606f73fef69beaa7014dccf63aa6793a0cc26746ac46ba0082e8e85"} Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.565603 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-590a-account-create-kxt2z" event={"ID":"81ec0e3a-1906-4013-a5cd-c639933a26a5","Type":"ContainerStarted","Data":"5cfa0c1f99d8f58d88da3e21538736b3254b18004ca9ba63fbb2098d4ce8b351"} Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.567059 4777 generic.go:334] "Generic (PLEG): container finished" podID="7c41309a-1061-4cf2-bf89-f170ee446905" containerID="bee6b663e803ef6ed74f656ae8ce0b476f0c81e8125c23b32c96b7fd47089a38" exitCode=0 Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.567113 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fzsq6" event={"ID":"7c41309a-1061-4cf2-bf89-f170ee446905","Type":"ContainerDied","Data":"bee6b663e803ef6ed74f656ae8ce0b476f0c81e8125c23b32c96b7fd47089a38"} Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.567171 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fzsq6" event={"ID":"7c41309a-1061-4cf2-bf89-f170ee446905","Type":"ContainerStarted","Data":"c45357784e3701fb6ab576d6be307236c575b14ee7c3e05ec38c2d11cb4f064f"} Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.573342 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"df409568-e0d5-49db-8da4-efbe45ec3a2c","Type":"ContainerStarted","Data":"35f69bd06934daf024dcc3109cb2ea14c9f384b51e4444d410222d50c3ea1551"} Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.577325 4777 generic.go:334] "Generic (PLEG): container finished" podID="c6389225-005e-4032-a3bb-a5700e55d08e" containerID="d279e716a60bd53020eb206467a900871b6af1b0724000a7ad1429e6114d7844" exitCode=0 Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.577394 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5d5f-account-create-vxk8c" event={"ID":"c6389225-005e-4032-a3bb-a5700e55d08e","Type":"ContainerDied","Data":"d279e716a60bd53020eb206467a900871b6af1b0724000a7ad1429e6114d7844"} Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.578625 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-stwcx" event={"ID":"cc393164-6269-4238-8eab-294157bd3e0a","Type":"ContainerStarted","Data":"69f63ca5ce9dbbc74352b7994c6c6c23bde8d9f0b087b35e6dfaf34541d769ff"} Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.579911 4777 generic.go:334] "Generic (PLEG): container finished" podID="e082ddb3-c5ea-4e65-b002-abefd538a16b" containerID="e52e01882ae7e79941c7f6b8a9a374979301f72a5efcf864645fa966abf14cca" exitCode=0 Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.579983 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-q9djb" event={"ID":"e082ddb3-c5ea-4e65-b002-abefd538a16b","Type":"ContainerDied","Data":"e52e01882ae7e79941c7f6b8a9a374979301f72a5efcf864645fa966abf14cca"} Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.581490 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-q4l2k-config-mzkfx" event={"ID":"7f6109d2-ed42-48f2-8998-f14ab220416f","Type":"ContainerDied","Data":"444b4970f3a3a0989db90b22c605d40b7297ba929a55b73028c1d8239c0c43fc"} Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.581513 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="444b4970f3a3a0989db90b22c605d40b7297ba929a55b73028c1d8239c0c43fc" Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.581543 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q4l2k-config-mzkfx" Nov 24 17:20:24 crc kubenswrapper[4777]: I1124 17:20:24.647940 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=17.647921909 podStartE2EDuration="17.647921909s" podCreationTimestamp="2025-11-24 17:20:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:20:24.635491421 +0000 UTC m=+1182.794686470" watchObservedRunningTime="2025-11-24 17:20:24.647921909 +0000 UTC m=+1182.807116958" Nov 24 17:20:25 crc kubenswrapper[4777]: I1124 17:20:25.159713 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-q4l2k-config-mzkfx"] Nov 24 17:20:25 crc kubenswrapper[4777]: I1124 17:20:25.175372 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-q4l2k-config-mzkfx"] Nov 24 17:20:25 crc kubenswrapper[4777]: I1124 17:20:25.263707 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f6109d2-ed42-48f2-8998-f14ab220416f" path="/var/lib/kubelet/pods/7f6109d2-ed42-48f2-8998-f14ab220416f/volumes" Nov 24 17:20:25 crc kubenswrapper[4777]: I1124 17:20:25.745912 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-q4l2k" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.396629 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5d5f-account-create-vxk8c" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.406773 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-q9djb" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.414793 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-590a-account-create-kxt2z" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.426716 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c352-account-create-r4cp5" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.453447 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fzsq6" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.458132 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-19af-account-create-9fl7d" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.460957 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlsxz\" (UniqueName: \"kubernetes.io/projected/c6389225-005e-4032-a3bb-a5700e55d08e-kube-api-access-zlsxz\") pod \"c6389225-005e-4032-a3bb-a5700e55d08e\" (UID: \"c6389225-005e-4032-a3bb-a5700e55d08e\") " Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.461056 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6389225-005e-4032-a3bb-a5700e55d08e-operator-scripts\") pod \"c6389225-005e-4032-a3bb-a5700e55d08e\" (UID: \"c6389225-005e-4032-a3bb-a5700e55d08e\") " Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.464078 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6389225-005e-4032-a3bb-a5700e55d08e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c6389225-005e-4032-a3bb-a5700e55d08e" (UID: "c6389225-005e-4032-a3bb-a5700e55d08e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.465256 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9666p" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.470346 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6389225-005e-4032-a3bb-a5700e55d08e-kube-api-access-zlsxz" (OuterVolumeSpecName: "kube-api-access-zlsxz") pod "c6389225-005e-4032-a3bb-a5700e55d08e" (UID: "c6389225-005e-4032-a3bb-a5700e55d08e"). InnerVolumeSpecName "kube-api-access-zlsxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.563542 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9721c23b-c04c-4dc4-94ab-24455982fbae-operator-scripts\") pod \"9721c23b-c04c-4dc4-94ab-24455982fbae\" (UID: \"9721c23b-c04c-4dc4-94ab-24455982fbae\") " Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.563688 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxgbk\" (UniqueName: \"kubernetes.io/projected/7c41309a-1061-4cf2-bf89-f170ee446905-kube-api-access-bxgbk\") pod \"7c41309a-1061-4cf2-bf89-f170ee446905\" (UID: \"7c41309a-1061-4cf2-bf89-f170ee446905\") " Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.563734 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d7c6a9d-cea0-4fab-abc8-53e1926587cc-operator-scripts\") pod \"8d7c6a9d-cea0-4fab-abc8-53e1926587cc\" (UID: \"8d7c6a9d-cea0-4fab-abc8-53e1926587cc\") " Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.563787 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c268k\" (UniqueName: \"kubernetes.io/projected/e082ddb3-c5ea-4e65-b002-abefd538a16b-kube-api-access-c268k\") pod \"e082ddb3-c5ea-4e65-b002-abefd538a16b\" (UID: \"e082ddb3-c5ea-4e65-b002-abefd538a16b\") " Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.563837 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c41309a-1061-4cf2-bf89-f170ee446905-operator-scripts\") pod \"7c41309a-1061-4cf2-bf89-f170ee446905\" (UID: \"7c41309a-1061-4cf2-bf89-f170ee446905\") " Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.563872 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e082ddb3-c5ea-4e65-b002-abefd538a16b-operator-scripts\") pod \"e082ddb3-c5ea-4e65-b002-abefd538a16b\" (UID: \"e082ddb3-c5ea-4e65-b002-abefd538a16b\") " Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.563899 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld6bb\" (UniqueName: \"kubernetes.io/projected/8d7c6a9d-cea0-4fab-abc8-53e1926587cc-kube-api-access-ld6bb\") pod \"8d7c6a9d-cea0-4fab-abc8-53e1926587cc\" (UID: \"8d7c6a9d-cea0-4fab-abc8-53e1926587cc\") " Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.563995 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81ec0e3a-1906-4013-a5cd-c639933a26a5-operator-scripts\") pod \"81ec0e3a-1906-4013-a5cd-c639933a26a5\" (UID: \"81ec0e3a-1906-4013-a5cd-c639933a26a5\") " Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.564015 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f527\" (UniqueName: \"kubernetes.io/projected/9721c23b-c04c-4dc4-94ab-24455982fbae-kube-api-access-2f527\") pod \"9721c23b-c04c-4dc4-94ab-24455982fbae\" (UID: \"9721c23b-c04c-4dc4-94ab-24455982fbae\") " Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.564095 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6jmx\" (UniqueName: \"kubernetes.io/projected/81ec0e3a-1906-4013-a5cd-c639933a26a5-kube-api-access-k6jmx\") pod \"81ec0e3a-1906-4013-a5cd-c639933a26a5\" (UID: \"81ec0e3a-1906-4013-a5cd-c639933a26a5\") " Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.564098 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9721c23b-c04c-4dc4-94ab-24455982fbae-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9721c23b-c04c-4dc4-94ab-24455982fbae" (UID: "9721c23b-c04c-4dc4-94ab-24455982fbae"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.564270 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab4c523b-a2f8-4c22-bbf8-c4ce606bd932-operator-scripts\") pod \"ab4c523b-a2f8-4c22-bbf8-c4ce606bd932\" (UID: \"ab4c523b-a2f8-4c22-bbf8-c4ce606bd932\") " Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.564348 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xk8f\" (UniqueName: \"kubernetes.io/projected/ab4c523b-a2f8-4c22-bbf8-c4ce606bd932-kube-api-access-5xk8f\") pod \"ab4c523b-a2f8-4c22-bbf8-c4ce606bd932\" (UID: \"ab4c523b-a2f8-4c22-bbf8-c4ce606bd932\") " Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.564392 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e082ddb3-c5ea-4e65-b002-abefd538a16b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e082ddb3-c5ea-4e65-b002-abefd538a16b" (UID: "e082ddb3-c5ea-4e65-b002-abefd538a16b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.565002 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7c6a9d-cea0-4fab-abc8-53e1926587cc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8d7c6a9d-cea0-4fab-abc8-53e1926587cc" (UID: "8d7c6a9d-cea0-4fab-abc8-53e1926587cc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.565080 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9721c23b-c04c-4dc4-94ab-24455982fbae-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.565100 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlsxz\" (UniqueName: \"kubernetes.io/projected/c6389225-005e-4032-a3bb-a5700e55d08e-kube-api-access-zlsxz\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.565117 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e082ddb3-c5ea-4e65-b002-abefd538a16b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.565130 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6389225-005e-4032-a3bb-a5700e55d08e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.565316 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab4c523b-a2f8-4c22-bbf8-c4ce606bd932-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ab4c523b-a2f8-4c22-bbf8-c4ce606bd932" (UID: "ab4c523b-a2f8-4c22-bbf8-c4ce606bd932"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.565363 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81ec0e3a-1906-4013-a5cd-c639933a26a5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "81ec0e3a-1906-4013-a5cd-c639933a26a5" (UID: "81ec0e3a-1906-4013-a5cd-c639933a26a5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.565392 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c41309a-1061-4cf2-bf89-f170ee446905-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7c41309a-1061-4cf2-bf89-f170ee446905" (UID: "7c41309a-1061-4cf2-bf89-f170ee446905"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.568187 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e082ddb3-c5ea-4e65-b002-abefd538a16b-kube-api-access-c268k" (OuterVolumeSpecName: "kube-api-access-c268k") pod "e082ddb3-c5ea-4e65-b002-abefd538a16b" (UID: "e082ddb3-c5ea-4e65-b002-abefd538a16b"). InnerVolumeSpecName "kube-api-access-c268k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.569313 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d7c6a9d-cea0-4fab-abc8-53e1926587cc-kube-api-access-ld6bb" (OuterVolumeSpecName: "kube-api-access-ld6bb") pod "8d7c6a9d-cea0-4fab-abc8-53e1926587cc" (UID: "8d7c6a9d-cea0-4fab-abc8-53e1926587cc"). InnerVolumeSpecName "kube-api-access-ld6bb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.569605 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81ec0e3a-1906-4013-a5cd-c639933a26a5-kube-api-access-k6jmx" (OuterVolumeSpecName: "kube-api-access-k6jmx") pod "81ec0e3a-1906-4013-a5cd-c639933a26a5" (UID: "81ec0e3a-1906-4013-a5cd-c639933a26a5"). InnerVolumeSpecName "kube-api-access-k6jmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.569704 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab4c523b-a2f8-4c22-bbf8-c4ce606bd932-kube-api-access-5xk8f" (OuterVolumeSpecName: "kube-api-access-5xk8f") pod "ab4c523b-a2f8-4c22-bbf8-c4ce606bd932" (UID: "ab4c523b-a2f8-4c22-bbf8-c4ce606bd932"). InnerVolumeSpecName "kube-api-access-5xk8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.570037 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9721c23b-c04c-4dc4-94ab-24455982fbae-kube-api-access-2f527" (OuterVolumeSpecName: "kube-api-access-2f527") pod "9721c23b-c04c-4dc4-94ab-24455982fbae" (UID: "9721c23b-c04c-4dc4-94ab-24455982fbae"). InnerVolumeSpecName "kube-api-access-2f527". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.581528 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c41309a-1061-4cf2-bf89-f170ee446905-kube-api-access-bxgbk" (OuterVolumeSpecName: "kube-api-access-bxgbk") pod "7c41309a-1061-4cf2-bf89-f170ee446905" (UID: "7c41309a-1061-4cf2-bf89-f170ee446905"). InnerVolumeSpecName "kube-api-access-bxgbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.612289 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c352-account-create-r4cp5" event={"ID":"ab4c523b-a2f8-4c22-bbf8-c4ce606bd932","Type":"ContainerDied","Data":"a2c4a94985c2354ea916dd320969d00bacec5083490e57c8c5f182932fd8947e"} Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.612330 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2c4a94985c2354ea916dd320969d00bacec5083490e57c8c5f182932fd8947e" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.612286 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c352-account-create-r4cp5" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.625429 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-q9djb" event={"ID":"e082ddb3-c5ea-4e65-b002-abefd538a16b","Type":"ContainerDied","Data":"3d2d4871f4112eb69eb0ebe7363b0be07a414a3dca8e02a640367921abaebefe"} Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.625479 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d2d4871f4112eb69eb0ebe7363b0be07a414a3dca8e02a640367921abaebefe" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.625569 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-q9djb" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.638756 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9666p" event={"ID":"8d7c6a9d-cea0-4fab-abc8-53e1926587cc","Type":"ContainerDied","Data":"9f09c3980739419e5631ad4050738c6180bb80a1ae4476e1a3e5d1879bd7274c"} Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.638807 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f09c3980739419e5631ad4050738c6180bb80a1ae4476e1a3e5d1879bd7274c" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.638939 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9666p" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.641668 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-590a-account-create-kxt2z" event={"ID":"81ec0e3a-1906-4013-a5cd-c639933a26a5","Type":"ContainerDied","Data":"5cfa0c1f99d8f58d88da3e21538736b3254b18004ca9ba63fbb2098d4ce8b351"} Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.641725 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cfa0c1f99d8f58d88da3e21538736b3254b18004ca9ba63fbb2098d4ce8b351" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.641767 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-590a-account-create-kxt2z" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.646486 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fzsq6" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.646454 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fzsq6" event={"ID":"7c41309a-1061-4cf2-bf89-f170ee446905","Type":"ContainerDied","Data":"c45357784e3701fb6ab576d6be307236c575b14ee7c3e05ec38c2d11cb4f064f"} Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.647610 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c45357784e3701fb6ab576d6be307236c575b14ee7c3e05ec38c2d11cb4f064f" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.650250 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-19af-account-create-9fl7d" event={"ID":"9721c23b-c04c-4dc4-94ab-24455982fbae","Type":"ContainerDied","Data":"eb2656291667de58817ba71399a143d59fef18221423408e7e52b6d285b62187"} Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.650289 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb2656291667de58817ba71399a143d59fef18221423408e7e52b6d285b62187" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.650315 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-19af-account-create-9fl7d" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.662705 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5d5f-account-create-vxk8c" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.662801 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5d5f-account-create-vxk8c" event={"ID":"c6389225-005e-4032-a3bb-a5700e55d08e","Type":"ContainerDied","Data":"1059575ea70673f1b2910ff87354c9f1af24c352a565c6f2732c2b173976a28e"} Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.662848 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1059575ea70673f1b2910ff87354c9f1af24c352a565c6f2732c2b173976a28e" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.667170 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xk8f\" (UniqueName: \"kubernetes.io/projected/ab4c523b-a2f8-4c22-bbf8-c4ce606bd932-kube-api-access-5xk8f\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.667198 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxgbk\" (UniqueName: \"kubernetes.io/projected/7c41309a-1061-4cf2-bf89-f170ee446905-kube-api-access-bxgbk\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.667210 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d7c6a9d-cea0-4fab-abc8-53e1926587cc-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.667219 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c268k\" (UniqueName: \"kubernetes.io/projected/e082ddb3-c5ea-4e65-b002-abefd538a16b-kube-api-access-c268k\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.667228 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c41309a-1061-4cf2-bf89-f170ee446905-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.667236 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld6bb\" (UniqueName: \"kubernetes.io/projected/8d7c6a9d-cea0-4fab-abc8-53e1926587cc-kube-api-access-ld6bb\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.667245 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81ec0e3a-1906-4013-a5cd-c639933a26a5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.667253 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2f527\" (UniqueName: \"kubernetes.io/projected/9721c23b-c04c-4dc4-94ab-24455982fbae-kube-api-access-2f527\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.667261 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6jmx\" (UniqueName: \"kubernetes.io/projected/81ec0e3a-1906-4013-a5cd-c639933a26a5-kube-api-access-k6jmx\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.667269 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab4c523b-a2f8-4c22-bbf8-c4ce606bd932-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:27 crc kubenswrapper[4777]: I1124 17:20:27.927607 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:30 crc kubenswrapper[4777]: I1124 17:20:30.696112 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-86vqg" Nov 24 17:20:30 crc kubenswrapper[4777]: I1124 17:20:30.702829 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-86vqg" event={"ID":"bf61c287-5b51-4b32-a50e-54b93835632d","Type":"ContainerDied","Data":"b60ed3b0c047c0d88dc42bc9b7a61e29fd33ca1e50e118645ca12b7d72a5befc"} Nov 24 17:20:30 crc kubenswrapper[4777]: I1124 17:20:30.702863 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b60ed3b0c047c0d88dc42bc9b7a61e29fd33ca1e50e118645ca12b7d72a5befc" Nov 24 17:20:30 crc kubenswrapper[4777]: I1124 17:20:30.702916 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-86vqg" Nov 24 17:20:30 crc kubenswrapper[4777]: I1124 17:20:30.842669 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khksx\" (UniqueName: \"kubernetes.io/projected/bf61c287-5b51-4b32-a50e-54b93835632d-kube-api-access-khksx\") pod \"bf61c287-5b51-4b32-a50e-54b93835632d\" (UID: \"bf61c287-5b51-4b32-a50e-54b93835632d\") " Nov 24 17:20:30 crc kubenswrapper[4777]: I1124 17:20:30.842813 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf61c287-5b51-4b32-a50e-54b93835632d-operator-scripts\") pod \"bf61c287-5b51-4b32-a50e-54b93835632d\" (UID: \"bf61c287-5b51-4b32-a50e-54b93835632d\") " Nov 24 17:20:30 crc kubenswrapper[4777]: I1124 17:20:30.843580 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf61c287-5b51-4b32-a50e-54b93835632d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bf61c287-5b51-4b32-a50e-54b93835632d" (UID: "bf61c287-5b51-4b32-a50e-54b93835632d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:30 crc kubenswrapper[4777]: I1124 17:20:30.851254 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf61c287-5b51-4b32-a50e-54b93835632d-kube-api-access-khksx" (OuterVolumeSpecName: "kube-api-access-khksx") pod "bf61c287-5b51-4b32-a50e-54b93835632d" (UID: "bf61c287-5b51-4b32-a50e-54b93835632d"). InnerVolumeSpecName "kube-api-access-khksx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:30 crc kubenswrapper[4777]: I1124 17:20:30.944981 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khksx\" (UniqueName: \"kubernetes.io/projected/bf61c287-5b51-4b32-a50e-54b93835632d-kube-api-access-khksx\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:30 crc kubenswrapper[4777]: I1124 17:20:30.945011 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf61c287-5b51-4b32-a50e-54b93835632d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:31 crc kubenswrapper[4777]: I1124 17:20:31.715942 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"452bf17e10606325a63f50c12d93a3f7e7eb807e324d98b04fc973605302a604"} Nov 24 17:20:31 crc kubenswrapper[4777]: I1124 17:20:31.716955 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"8e00b946862d56e499185192f623f9d53f087e0d9065c4717ced79702534592b"} Nov 24 17:20:31 crc kubenswrapper[4777]: I1124 17:20:31.716989 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"34d3c556f3d78552af3cb5a0361b9f3e7a203b78677b6aea613197385ef8de08"} Nov 24 17:20:31 crc kubenswrapper[4777]: I1124 17:20:31.716999 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"aa82bfdce0ef318d7045aa6cd2d4042cdcb9eb17c0a380569c549e71fe005057"} Nov 24 17:20:31 crc kubenswrapper[4777]: I1124 17:20:31.717857 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-stwcx" event={"ID":"cc393164-6269-4238-8eab-294157bd3e0a","Type":"ContainerStarted","Data":"57063a5842a346b840a1e19cfdac14c41890e1df1a569da8e9888af51709f7fa"} Nov 24 17:20:31 crc kubenswrapper[4777]: I1124 17:20:31.745094 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-stwcx" podStartSLOduration=3.060250884 podStartE2EDuration="9.745070453s" podCreationTimestamp="2025-11-24 17:20:22 +0000 UTC" firstStartedPulling="2025-11-24 17:20:23.76633839 +0000 UTC m=+1181.925533439" lastFinishedPulling="2025-11-24 17:20:30.451157959 +0000 UTC m=+1188.610353008" observedRunningTime="2025-11-24 17:20:31.734057007 +0000 UTC m=+1189.893252056" watchObservedRunningTime="2025-11-24 17:20:31.745070453 +0000 UTC m=+1189.904265502" Nov 24 17:20:33 crc kubenswrapper[4777]: I1124 17:20:33.738816 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"179d0e7c0b034e07fc2082abf122475bbb7cf3bd1484bac7c264d01a12c70bac"} Nov 24 17:20:33 crc kubenswrapper[4777]: I1124 17:20:33.739369 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"0527ad06c56f7d5c50f7e4bad810f4b5191b032146c01ca4ce8c0f183511de83"} Nov 24 17:20:34 crc kubenswrapper[4777]: I1124 17:20:34.758531 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"6f34bf4ebf0306b1ae1684f47525d903f042c42ebe9ba6fb794a0c8ad1435b5d"} Nov 24 17:20:34 crc kubenswrapper[4777]: I1124 17:20:34.758574 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"519b33b03cc438dc1febf283c6ab87bafc33f15a350ed45d940bb82adc191534"} Nov 24 17:20:35 crc kubenswrapper[4777]: I1124 17:20:35.787846 4777 generic.go:334] "Generic (PLEG): container finished" podID="7b60b793-78a5-406c-9ed6-e582dfba568b" containerID="d57f5c5e958e3c77e35416bfd8f87db04e97a311ca0c6689b7decfa910bf7618" exitCode=0 Nov 24 17:20:35 crc kubenswrapper[4777]: I1124 17:20:35.788325 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-swrgj" event={"ID":"7b60b793-78a5-406c-9ed6-e582dfba568b","Type":"ContainerDied","Data":"d57f5c5e958e3c77e35416bfd8f87db04e97a311ca0c6689b7decfa910bf7618"} Nov 24 17:20:35 crc kubenswrapper[4777]: I1124 17:20:35.827899 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"60ac84a86410f070427239d1a4f910b312a777f0306062c002b537b729c46e84"} Nov 24 17:20:36 crc kubenswrapper[4777]: I1124 17:20:36.841121 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"f78ff423d56f8409657d24b49aa8aaef62485019a6c6fe6fc7fdab8ee62abb14"} Nov 24 17:20:36 crc kubenswrapper[4777]: I1124 17:20:36.841569 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"2f2456eb56763d43b520d7efb672cd2034792e166310c830c99f14c8f9e6dd1d"} Nov 24 17:20:36 crc kubenswrapper[4777]: I1124 17:20:36.841586 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"239479b283ae44ba34b1790f355da4718d6c590a2594ff7139597d819b3d0caf"} Nov 24 17:20:36 crc kubenswrapper[4777]: I1124 17:20:36.841598 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"236aa3fac444d667d835871384795571bb224ccaf1f8841e217affe24076b4b1"} Nov 24 17:20:36 crc kubenswrapper[4777]: I1124 17:20:36.845118 4777 generic.go:334] "Generic (PLEG): container finished" podID="cc393164-6269-4238-8eab-294157bd3e0a" containerID="57063a5842a346b840a1e19cfdac14c41890e1df1a569da8e9888af51709f7fa" exitCode=0 Nov 24 17:20:36 crc kubenswrapper[4777]: I1124 17:20:36.845192 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-stwcx" event={"ID":"cc393164-6269-4238-8eab-294157bd3e0a","Type":"ContainerDied","Data":"57063a5842a346b840a1e19cfdac14c41890e1df1a569da8e9888af51709f7fa"} Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.310192 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.467904 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-db-sync-config-data\") pod \"7b60b793-78a5-406c-9ed6-e582dfba568b\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.467952 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmwh6\" (UniqueName: \"kubernetes.io/projected/7b60b793-78a5-406c-9ed6-e582dfba568b-kube-api-access-tmwh6\") pod \"7b60b793-78a5-406c-9ed6-e582dfba568b\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.468258 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-config-data\") pod \"7b60b793-78a5-406c-9ed6-e582dfba568b\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.468304 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-combined-ca-bundle\") pod \"7b60b793-78a5-406c-9ed6-e582dfba568b\" (UID: \"7b60b793-78a5-406c-9ed6-e582dfba568b\") " Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.476262 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7b60b793-78a5-406c-9ed6-e582dfba568b" (UID: "7b60b793-78a5-406c-9ed6-e582dfba568b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.477024 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b60b793-78a5-406c-9ed6-e582dfba568b-kube-api-access-tmwh6" (OuterVolumeSpecName: "kube-api-access-tmwh6") pod "7b60b793-78a5-406c-9ed6-e582dfba568b" (UID: "7b60b793-78a5-406c-9ed6-e582dfba568b"). InnerVolumeSpecName "kube-api-access-tmwh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.498167 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b60b793-78a5-406c-9ed6-e582dfba568b" (UID: "7b60b793-78a5-406c-9ed6-e582dfba568b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.523254 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-config-data" (OuterVolumeSpecName: "config-data") pod "7b60b793-78a5-406c-9ed6-e582dfba568b" (UID: "7b60b793-78a5-406c-9ed6-e582dfba568b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.570568 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.570605 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.570619 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmwh6\" (UniqueName: \"kubernetes.io/projected/7b60b793-78a5-406c-9ed6-e582dfba568b-kube-api-access-tmwh6\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.570630 4777 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7b60b793-78a5-406c-9ed6-e582dfba568b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.858954 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-swrgj" Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.858958 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-swrgj" event={"ID":"7b60b793-78a5-406c-9ed6-e582dfba568b","Type":"ContainerDied","Data":"f827e21383ff8622f7cfba44697eb036fa2556d1b57b579fc0bc0dcd63d10c1c"} Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.859031 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f827e21383ff8622f7cfba44697eb036fa2556d1b57b579fc0bc0dcd63d10c1c" Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.864646 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"bb04fb78e34f5e9c02ad5be27c22fa89192aea12f3401a39a20f06d5b8fafa51"} Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.864696 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9c412d58-6198-4254-932d-5eb90420f8ba","Type":"ContainerStarted","Data":"d50e009a9b02e45ca31190e98b3cca96faa754141fa42cea559da8eab3712880"} Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.920272 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=69.495646844 podStartE2EDuration="1m21.920245067s" podCreationTimestamp="2025-11-24 17:19:16 +0000 UTC" firstStartedPulling="2025-11-24 17:20:23.059400155 +0000 UTC m=+1181.218595204" lastFinishedPulling="2025-11-24 17:20:35.483998378 +0000 UTC m=+1193.643193427" observedRunningTime="2025-11-24 17:20:37.910781415 +0000 UTC m=+1196.069976494" watchObservedRunningTime="2025-11-24 17:20:37.920245067 +0000 UTC m=+1196.079440126" Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.931753 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:37 crc kubenswrapper[4777]: I1124 17:20:37.980120 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.265489 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-85lnk"] Nov 24 17:20:38 crc kubenswrapper[4777]: E1124 17:20:38.266368 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f6109d2-ed42-48f2-8998-f14ab220416f" containerName="ovn-config" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266396 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f6109d2-ed42-48f2-8998-f14ab220416f" containerName="ovn-config" Nov 24 17:20:38 crc kubenswrapper[4777]: E1124 17:20:38.266409 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e082ddb3-c5ea-4e65-b002-abefd538a16b" containerName="mariadb-database-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266417 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="e082ddb3-c5ea-4e65-b002-abefd538a16b" containerName="mariadb-database-create" Nov 24 17:20:38 crc kubenswrapper[4777]: E1124 17:20:38.266441 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6389225-005e-4032-a3bb-a5700e55d08e" containerName="mariadb-account-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266449 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6389225-005e-4032-a3bb-a5700e55d08e" containerName="mariadb-account-create" Nov 24 17:20:38 crc kubenswrapper[4777]: E1124 17:20:38.266459 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf61c287-5b51-4b32-a50e-54b93835632d" containerName="mariadb-database-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266467 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf61c287-5b51-4b32-a50e-54b93835632d" containerName="mariadb-database-create" Nov 24 17:20:38 crc kubenswrapper[4777]: E1124 17:20:38.266483 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c41309a-1061-4cf2-bf89-f170ee446905" containerName="mariadb-database-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266492 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c41309a-1061-4cf2-bf89-f170ee446905" containerName="mariadb-database-create" Nov 24 17:20:38 crc kubenswrapper[4777]: E1124 17:20:38.266502 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81ec0e3a-1906-4013-a5cd-c639933a26a5" containerName="mariadb-account-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266510 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="81ec0e3a-1906-4013-a5cd-c639933a26a5" containerName="mariadb-account-create" Nov 24 17:20:38 crc kubenswrapper[4777]: E1124 17:20:38.266522 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7c6a9d-cea0-4fab-abc8-53e1926587cc" containerName="mariadb-database-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266529 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7c6a9d-cea0-4fab-abc8-53e1926587cc" containerName="mariadb-database-create" Nov 24 17:20:38 crc kubenswrapper[4777]: E1124 17:20:38.266547 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9721c23b-c04c-4dc4-94ab-24455982fbae" containerName="mariadb-account-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266554 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9721c23b-c04c-4dc4-94ab-24455982fbae" containerName="mariadb-account-create" Nov 24 17:20:38 crc kubenswrapper[4777]: E1124 17:20:38.266591 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b60b793-78a5-406c-9ed6-e582dfba568b" containerName="glance-db-sync" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266599 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b60b793-78a5-406c-9ed6-e582dfba568b" containerName="glance-db-sync" Nov 24 17:20:38 crc kubenswrapper[4777]: E1124 17:20:38.266613 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab4c523b-a2f8-4c22-bbf8-c4ce606bd932" containerName="mariadb-account-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266632 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab4c523b-a2f8-4c22-bbf8-c4ce606bd932" containerName="mariadb-account-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266843 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f6109d2-ed42-48f2-8998-f14ab220416f" containerName="ovn-config" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266871 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="e082ddb3-c5ea-4e65-b002-abefd538a16b" containerName="mariadb-database-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266886 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b60b793-78a5-406c-9ed6-e582dfba568b" containerName="glance-db-sync" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266897 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9721c23b-c04c-4dc4-94ab-24455982fbae" containerName="mariadb-account-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266911 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d7c6a9d-cea0-4fab-abc8-53e1926587cc" containerName="mariadb-database-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266923 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab4c523b-a2f8-4c22-bbf8-c4ce606bd932" containerName="mariadb-account-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266937 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf61c287-5b51-4b32-a50e-54b93835632d" containerName="mariadb-database-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266948 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c41309a-1061-4cf2-bf89-f170ee446905" containerName="mariadb-database-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.266985 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="81ec0e3a-1906-4013-a5cd-c639933a26a5" containerName="mariadb-account-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.267000 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6389225-005e-4032-a3bb-a5700e55d08e" containerName="mariadb-account-create" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.292473 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-85lnk"] Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.292589 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.352181 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-85lnk"] Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.380452 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-c78nd"] Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.384726 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.393334 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-85lnk\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.393469 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-85lnk\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.393499 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-config\") pod \"dnsmasq-dns-74dc88fc-85lnk\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.393551 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dnbj\" (UniqueName: \"kubernetes.io/projected/25f20c17-1340-4d4e-897a-8ca3230d44b8-kube-api-access-9dnbj\") pod \"dnsmasq-dns-74dc88fc-85lnk\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.393591 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-dns-svc\") pod \"dnsmasq-dns-74dc88fc-85lnk\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.395280 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.402539 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-c78nd"] Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.418711 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-stwcx" Nov 24 17:20:38 crc kubenswrapper[4777]: E1124 17:20:38.446151 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-9dnbj ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-74dc88fc-85lnk" podUID="25f20c17-1340-4d4e-897a-8ca3230d44b8" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.495441 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-492bz\" (UniqueName: \"kubernetes.io/projected/cc393164-6269-4238-8eab-294157bd3e0a-kube-api-access-492bz\") pod \"cc393164-6269-4238-8eab-294157bd3e0a\" (UID: \"cc393164-6269-4238-8eab-294157bd3e0a\") " Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.495766 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc393164-6269-4238-8eab-294157bd3e0a-config-data\") pod \"cc393164-6269-4238-8eab-294157bd3e0a\" (UID: \"cc393164-6269-4238-8eab-294157bd3e0a\") " Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.495884 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc393164-6269-4238-8eab-294157bd3e0a-combined-ca-bundle\") pod \"cc393164-6269-4238-8eab-294157bd3e0a\" (UID: \"cc393164-6269-4238-8eab-294157bd3e0a\") " Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.496282 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-config\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.496427 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-85lnk\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.496528 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-config\") pod \"dnsmasq-dns-74dc88fc-85lnk\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.496617 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.496715 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dnbj\" (UniqueName: \"kubernetes.io/projected/25f20c17-1340-4d4e-897a-8ca3230d44b8-kube-api-access-9dnbj\") pod \"dnsmasq-dns-74dc88fc-85lnk\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.496800 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.496957 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-dns-svc\") pod \"dnsmasq-dns-74dc88fc-85lnk\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.497080 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.497159 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99kn7\" (UniqueName: \"kubernetes.io/projected/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-kube-api-access-99kn7\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.497245 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.497325 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-85lnk\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.497890 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-85lnk\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.498815 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-config\") pod \"dnsmasq-dns-74dc88fc-85lnk\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.498994 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc393164-6269-4238-8eab-294157bd3e0a-kube-api-access-492bz" (OuterVolumeSpecName: "kube-api-access-492bz") pod "cc393164-6269-4238-8eab-294157bd3e0a" (UID: "cc393164-6269-4238-8eab-294157bd3e0a"). InnerVolumeSpecName "kube-api-access-492bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.499484 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-dns-svc\") pod \"dnsmasq-dns-74dc88fc-85lnk\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.501573 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-85lnk\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.516592 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dnbj\" (UniqueName: \"kubernetes.io/projected/25f20c17-1340-4d4e-897a-8ca3230d44b8-kube-api-access-9dnbj\") pod \"dnsmasq-dns-74dc88fc-85lnk\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.519301 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc393164-6269-4238-8eab-294157bd3e0a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc393164-6269-4238-8eab-294157bd3e0a" (UID: "cc393164-6269-4238-8eab-294157bd3e0a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.543625 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc393164-6269-4238-8eab-294157bd3e0a-config-data" (OuterVolumeSpecName: "config-data") pod "cc393164-6269-4238-8eab-294157bd3e0a" (UID: "cc393164-6269-4238-8eab-294157bd3e0a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.599400 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.599458 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.599513 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.599532 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99kn7\" (UniqueName: \"kubernetes.io/projected/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-kube-api-access-99kn7\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.599551 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.599589 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-config\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.599667 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc393164-6269-4238-8eab-294157bd3e0a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.599681 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc393164-6269-4238-8eab-294157bd3e0a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.599691 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-492bz\" (UniqueName: \"kubernetes.io/projected/cc393164-6269-4238-8eab-294157bd3e0a-kube-api-access-492bz\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.601037 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-config\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.601711 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.602486 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.603059 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.603562 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.618189 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99kn7\" (UniqueName: \"kubernetes.io/projected/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-kube-api-access-99kn7\") pod \"dnsmasq-dns-5f59b8f679-c78nd\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.763206 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.941112 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.941922 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-stwcx" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.953164 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-stwcx" event={"ID":"cc393164-6269-4238-8eab-294157bd3e0a","Type":"ContainerDied","Data":"69f63ca5ce9dbbc74352b7994c6c6c23bde8d9f0b087b35e6dfaf34541d769ff"} Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.953209 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69f63ca5ce9dbbc74352b7994c6c6c23bde8d9f0b087b35e6dfaf34541d769ff" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.976420 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:38 crc kubenswrapper[4777]: I1124 17:20:38.981464 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.073258 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-c78nd"] Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.118077 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dnbj\" (UniqueName: \"kubernetes.io/projected/25f20c17-1340-4d4e-897a-8ca3230d44b8-kube-api-access-9dnbj\") pod \"25f20c17-1340-4d4e-897a-8ca3230d44b8\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.118216 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-ovsdbserver-nb\") pod \"25f20c17-1340-4d4e-897a-8ca3230d44b8\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.118278 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-dns-svc\") pod \"25f20c17-1340-4d4e-897a-8ca3230d44b8\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.118358 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-ovsdbserver-sb\") pod \"25f20c17-1340-4d4e-897a-8ca3230d44b8\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.118429 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-config\") pod \"25f20c17-1340-4d4e-897a-8ca3230d44b8\" (UID: \"25f20c17-1340-4d4e-897a-8ca3230d44b8\") " Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.120421 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "25f20c17-1340-4d4e-897a-8ca3230d44b8" (UID: "25f20c17-1340-4d4e-897a-8ca3230d44b8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.120703 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "25f20c17-1340-4d4e-897a-8ca3230d44b8" (UID: "25f20c17-1340-4d4e-897a-8ca3230d44b8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.120912 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "25f20c17-1340-4d4e-897a-8ca3230d44b8" (UID: "25f20c17-1340-4d4e-897a-8ca3230d44b8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.121184 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-s2227"] Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.121538 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-config" (OuterVolumeSpecName: "config") pod "25f20c17-1340-4d4e-897a-8ca3230d44b8" (UID: "25f20c17-1340-4d4e-897a-8ca3230d44b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:39 crc kubenswrapper[4777]: E1124 17:20:39.121541 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc393164-6269-4238-8eab-294157bd3e0a" containerName="keystone-db-sync" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.121574 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc393164-6269-4238-8eab-294157bd3e0a" containerName="keystone-db-sync" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.121922 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc393164-6269-4238-8eab-294157bd3e0a" containerName="keystone-db-sync" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.122586 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.126550 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.126809 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.127019 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vzrxj" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.127133 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.127236 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.129435 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-cxtnv"] Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.131466 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.165925 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-s2227"] Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.169803 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25f20c17-1340-4d4e-897a-8ca3230d44b8-kube-api-access-9dnbj" (OuterVolumeSpecName: "kube-api-access-9dnbj") pod "25f20c17-1340-4d4e-897a-8ca3230d44b8" (UID: "25f20c17-1340-4d4e-897a-8ca3230d44b8"). InnerVolumeSpecName "kube-api-access-9dnbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.197280 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-cxtnv"] Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.222956 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-credential-keys\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223070 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-config\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223089 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223112 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223159 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-combined-ca-bundle\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223178 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwl56\" (UniqueName: \"kubernetes.io/projected/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-kube-api-access-gwl56\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223229 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-fernet-keys\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223258 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-scripts\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223286 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223308 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223329 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-config-data\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223377 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l66n8\" (UniqueName: \"kubernetes.io/projected/62219b70-7aac-4e82-9296-371a2283b6b0-kube-api-access-l66n8\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223423 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223433 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dnbj\" (UniqueName: \"kubernetes.io/projected/25f20c17-1340-4d4e-897a-8ca3230d44b8-kube-api-access-9dnbj\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223443 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223451 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.223459 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/25f20c17-1340-4d4e-897a-8ca3230d44b8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.325180 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-credential-keys\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.325244 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-config\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.325271 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.325308 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.325361 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-combined-ca-bundle\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.325383 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwl56\" (UniqueName: \"kubernetes.io/projected/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-kube-api-access-gwl56\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.325462 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-fernet-keys\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.325500 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-scripts\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.325528 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.325559 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.325583 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-config-data\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.325643 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l66n8\" (UniqueName: \"kubernetes.io/projected/62219b70-7aac-4e82-9296-371a2283b6b0-kube-api-access-l66n8\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.334524 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.334597 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-credential-keys\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.335239 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.336103 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.336250 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-config\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.337192 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.338253 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-scripts\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.339728 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-c78nd"] Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.340477 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-fernet-keys\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.348381 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-config-data\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.369152 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-combined-ca-bundle\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.383669 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l66n8\" (UniqueName: \"kubernetes.io/projected/62219b70-7aac-4e82-9296-371a2283b6b0-kube-api-access-l66n8\") pod \"keystone-bootstrap-s2227\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.389256 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwl56\" (UniqueName: \"kubernetes.io/projected/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-kube-api-access-gwl56\") pod \"dnsmasq-dns-bbf5cc879-cxtnv\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.454729 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.455444 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.457009 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.477321 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.478038 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.478353 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.509609 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.511638 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-kl4mk"] Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.515589 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kl4mk" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.531106 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.531144 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.531176 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0eeb390-6ffc-41c8-a816-0b4454730e07-run-httpd\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.531192 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-config-data\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.531225 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd7sx\" (UniqueName: \"kubernetes.io/projected/c0eeb390-6ffc-41c8-a816-0b4454730e07-kube-api-access-qd7sx\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.531249 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-scripts\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.531291 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0eeb390-6ffc-41c8-a816-0b4454730e07-log-httpd\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.544074 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-j8kfp" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.544478 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.544669 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.553622 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kl4mk"] Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.596387 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-sync-57hm5"] Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.597682 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.602862 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-l2r9w" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.615640 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.615742 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.615835 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.645108 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b01f3e72-974a-4649-84a3-d23609d774a9-config\") pod \"neutron-db-sync-kl4mk\" (UID: \"b01f3e72-974a-4649-84a3-d23609d774a9\") " pod="openstack/neutron-db-sync-kl4mk" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.645145 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-config-data\") pod \"cloudkitty-db-sync-57hm5\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.645193 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.645214 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.645241 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0eeb390-6ffc-41c8-a816-0b4454730e07-run-httpd\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.645258 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-config-data\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.645295 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd7sx\" (UniqueName: \"kubernetes.io/projected/c0eeb390-6ffc-41c8-a816-0b4454730e07-kube-api-access-qd7sx\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.645320 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zszjk\" (UniqueName: \"kubernetes.io/projected/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-kube-api-access-zszjk\") pod \"cloudkitty-db-sync-57hm5\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.645340 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-scripts\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.645382 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsv5x\" (UniqueName: \"kubernetes.io/projected/b01f3e72-974a-4649-84a3-d23609d774a9-kube-api-access-dsv5x\") pod \"neutron-db-sync-kl4mk\" (UID: \"b01f3e72-974a-4649-84a3-d23609d774a9\") " pod="openstack/neutron-db-sync-kl4mk" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.645401 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-certs\") pod \"cloudkitty-db-sync-57hm5\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.645419 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0eeb390-6ffc-41c8-a816-0b4454730e07-log-httpd\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.645433 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-combined-ca-bundle\") pod \"cloudkitty-db-sync-57hm5\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.645456 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-scripts\") pod \"cloudkitty-db-sync-57hm5\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.645476 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b01f3e72-974a-4649-84a3-d23609d774a9-combined-ca-bundle\") pod \"neutron-db-sync-kl4mk\" (UID: \"b01f3e72-974a-4649-84a3-d23609d774a9\") " pod="openstack/neutron-db-sync-kl4mk" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.646847 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0eeb390-6ffc-41c8-a816-0b4454730e07-log-httpd\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.654046 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0eeb390-6ffc-41c8-a816-0b4454730e07-run-httpd\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.723014 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-57hm5"] Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.747926 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsv5x\" (UniqueName: \"kubernetes.io/projected/b01f3e72-974a-4649-84a3-d23609d774a9-kube-api-access-dsv5x\") pod \"neutron-db-sync-kl4mk\" (UID: \"b01f3e72-974a-4649-84a3-d23609d774a9\") " pod="openstack/neutron-db-sync-kl4mk" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.747996 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-certs\") pod \"cloudkitty-db-sync-57hm5\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.748021 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-combined-ca-bundle\") pod \"cloudkitty-db-sync-57hm5\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.748041 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-scripts\") pod \"cloudkitty-db-sync-57hm5\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.748065 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b01f3e72-974a-4649-84a3-d23609d774a9-combined-ca-bundle\") pod \"neutron-db-sync-kl4mk\" (UID: \"b01f3e72-974a-4649-84a3-d23609d774a9\") " pod="openstack/neutron-db-sync-kl4mk" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.748103 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b01f3e72-974a-4649-84a3-d23609d774a9-config\") pod \"neutron-db-sync-kl4mk\" (UID: \"b01f3e72-974a-4649-84a3-d23609d774a9\") " pod="openstack/neutron-db-sync-kl4mk" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.748117 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-config-data\") pod \"cloudkitty-db-sync-57hm5\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.748214 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zszjk\" (UniqueName: \"kubernetes.io/projected/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-kube-api-access-zszjk\") pod \"cloudkitty-db-sync-57hm5\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.877391 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.877467 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b01f3e72-974a-4649-84a3-d23609d774a9-combined-ca-bundle\") pod \"neutron-db-sync-kl4mk\" (UID: \"b01f3e72-974a-4649-84a3-d23609d774a9\") " pod="openstack/neutron-db-sync-kl4mk" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.877916 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-scripts\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.878224 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.881632 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b01f3e72-974a-4649-84a3-d23609d774a9-config\") pod \"neutron-db-sync-kl4mk\" (UID: \"b01f3e72-974a-4649-84a3-d23609d774a9\") " pod="openstack/neutron-db-sync-kl4mk" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.882127 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-certs\") pod \"cloudkitty-db-sync-57hm5\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.883591 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd7sx\" (UniqueName: \"kubernetes.io/projected/c0eeb390-6ffc-41c8-a816-0b4454730e07-kube-api-access-qd7sx\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.885521 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-config-data\") pod \"ceilometer-0\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " pod="openstack/ceilometer-0" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.892197 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-config-data\") pod \"cloudkitty-db-sync-57hm5\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.892563 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-scripts\") pod \"cloudkitty-db-sync-57hm5\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.893711 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-combined-ca-bundle\") pod \"cloudkitty-db-sync-57hm5\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.905083 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-6wwt4"] Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.907005 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zszjk\" (UniqueName: \"kubernetes.io/projected/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-kube-api-access-zszjk\") pod \"cloudkitty-db-sync-57hm5\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.908755 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.909588 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsv5x\" (UniqueName: \"kubernetes.io/projected/b01f3e72-974a-4649-84a3-d23609d774a9-kube-api-access-dsv5x\") pod \"neutron-db-sync-kl4mk\" (UID: \"b01f3e72-974a-4649-84a3-d23609d774a9\") " pod="openstack/neutron-db-sync-kl4mk" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.941454 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9kzgd" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.941642 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.954130 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.955975 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-etc-machine-id\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.956061 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-combined-ca-bundle\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.956100 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-db-sync-config-data\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.956416 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l8gn\" (UniqueName: \"kubernetes.io/projected/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-kube-api-access-5l8gn\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.956484 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-scripts\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.956544 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-config-data\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.964072 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-v7hr7"] Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.965319 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-v7hr7" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.971094 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-6wwt4"] Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.989560 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-85lnk" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.990525 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" event={"ID":"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df","Type":"ContainerStarted","Data":"249ebbfc1b122cd30d90b79ba81aa9899a32526f5e730672e8cda507429ad4d2"} Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.993751 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 17:20:39 crc kubenswrapper[4777]: I1124 17:20:39.994260 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-f5rnw" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.004292 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-v7hr7"] Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.004733 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.031025 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-cxtnv"] Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.078036 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-db-sync-config-data\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.078120 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4151db5-e476-4ffa-a038-369874b439cc-combined-ca-bundle\") pod \"barbican-db-sync-v7hr7\" (UID: \"d4151db5-e476-4ffa-a038-369874b439cc\") " pod="openstack/barbican-db-sync-v7hr7" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.078152 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rszcl\" (UniqueName: \"kubernetes.io/projected/d4151db5-e476-4ffa-a038-369874b439cc-kube-api-access-rszcl\") pod \"barbican-db-sync-v7hr7\" (UID: \"d4151db5-e476-4ffa-a038-369874b439cc\") " pod="openstack/barbican-db-sync-v7hr7" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.078297 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l8gn\" (UniqueName: \"kubernetes.io/projected/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-kube-api-access-5l8gn\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.078324 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d4151db5-e476-4ffa-a038-369874b439cc-db-sync-config-data\") pod \"barbican-db-sync-v7hr7\" (UID: \"d4151db5-e476-4ffa-a038-369874b439cc\") " pod="openstack/barbican-db-sync-v7hr7" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.078345 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-scripts\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.078383 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-config-data\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.078429 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-etc-machine-id\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.078450 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-combined-ca-bundle\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.083467 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-db-sync-config-data\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.087148 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-scripts\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.089631 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-etc-machine-id\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.090504 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-config-data\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.100857 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-combined-ca-bundle\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.115532 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.128548 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-pwl4w"] Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.133871 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.138640 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.138958 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-44v99" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.140223 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.171688 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l8gn\" (UniqueName: \"kubernetes.io/projected/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-kube-api-access-5l8gn\") pod \"cinder-db-sync-6wwt4\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.172648 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kl4mk" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.201497 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-pwl4w"] Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.209435 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d84e561b-fba7-4438-9959-18e1b4bd889c-logs\") pod \"placement-db-sync-pwl4w\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.209502 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-combined-ca-bundle\") pod \"placement-db-sync-pwl4w\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.209558 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4151db5-e476-4ffa-a038-369874b439cc-combined-ca-bundle\") pod \"barbican-db-sync-v7hr7\" (UID: \"d4151db5-e476-4ffa-a038-369874b439cc\") " pod="openstack/barbican-db-sync-v7hr7" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.209587 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rszcl\" (UniqueName: \"kubernetes.io/projected/d4151db5-e476-4ffa-a038-369874b439cc-kube-api-access-rszcl\") pod \"barbican-db-sync-v7hr7\" (UID: \"d4151db5-e476-4ffa-a038-369874b439cc\") " pod="openstack/barbican-db-sync-v7hr7" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.209607 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-config-data\") pod \"placement-db-sync-pwl4w\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.209655 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-scripts\") pod \"placement-db-sync-pwl4w\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.209684 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h27lh\" (UniqueName: \"kubernetes.io/projected/d84e561b-fba7-4438-9959-18e1b4bd889c-kube-api-access-h27lh\") pod \"placement-db-sync-pwl4w\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.209746 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d4151db5-e476-4ffa-a038-369874b439cc-db-sync-config-data\") pod \"barbican-db-sync-v7hr7\" (UID: \"d4151db5-e476-4ffa-a038-369874b439cc\") " pod="openstack/barbican-db-sync-v7hr7" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.216448 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4151db5-e476-4ffa-a038-369874b439cc-combined-ca-bundle\") pod \"barbican-db-sync-v7hr7\" (UID: \"d4151db5-e476-4ffa-a038-369874b439cc\") " pod="openstack/barbican-db-sync-v7hr7" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.225385 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d4151db5-e476-4ffa-a038-369874b439cc-db-sync-config-data\") pod \"barbican-db-sync-v7hr7\" (UID: \"d4151db5-e476-4ffa-a038-369874b439cc\") " pod="openstack/barbican-db-sync-v7hr7" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.250146 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rszcl\" (UniqueName: \"kubernetes.io/projected/d4151db5-e476-4ffa-a038-369874b439cc-kube-api-access-rszcl\") pod \"barbican-db-sync-v7hr7\" (UID: \"d4151db5-e476-4ffa-a038-369874b439cc\") " pod="openstack/barbican-db-sync-v7hr7" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.252956 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-84jqq"] Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.254633 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.266794 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.285524 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-v7hr7" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.312008 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-scripts\") pod \"placement-db-sync-pwl4w\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.312077 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h27lh\" (UniqueName: \"kubernetes.io/projected/d84e561b-fba7-4438-9959-18e1b4bd889c-kube-api-access-h27lh\") pod \"placement-db-sync-pwl4w\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.312174 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d84e561b-fba7-4438-9959-18e1b4bd889c-logs\") pod \"placement-db-sync-pwl4w\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.312221 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-combined-ca-bundle\") pod \"placement-db-sync-pwl4w\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.312263 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-config-data\") pod \"placement-db-sync-pwl4w\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.317368 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-config-data\") pod \"placement-db-sync-pwl4w\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.322016 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d84e561b-fba7-4438-9959-18e1b4bd889c-logs\") pod \"placement-db-sync-pwl4w\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.323382 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-84jqq"] Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.350381 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-scripts\") pod \"placement-db-sync-pwl4w\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.351424 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-combined-ca-bundle\") pod \"placement-db-sync-pwl4w\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.360100 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h27lh\" (UniqueName: \"kubernetes.io/projected/d84e561b-fba7-4438-9959-18e1b4bd889c-kube-api-access-h27lh\") pod \"placement-db-sync-pwl4w\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.424340 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.425012 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.424373 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-85lnk"] Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.425271 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.425437 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8f9s\" (UniqueName: \"kubernetes.io/projected/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-kube-api-access-s8f9s\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.425679 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-config\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.429062 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.435632 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-85lnk"] Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.460899 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.462641 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.466937 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.467611 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-v86dz" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.467756 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.470070 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.527457 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.529251 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.530132 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.530186 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8f9s\" (UniqueName: \"kubernetes.io/projected/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-kube-api-access-s8f9s\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.530276 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-config\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.530293 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.530318 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.530333 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.531347 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.531433 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.534713 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-config\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.534713 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.557589 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-cxtnv"] Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.562596 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.563243 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.569786 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8f9s\" (UniqueName: \"kubernetes.io/projected/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-kube-api-access-s8f9s\") pod \"dnsmasq-dns-56df8fb6b7-84jqq\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.570597 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-s2227"] Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.584252 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:20:40 crc kubenswrapper[4777]: W1124 17:20:40.587798 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62219b70_7aac_4e82_9296_371a2283b6b0.slice/crio-00d5af6e019612533482017f0604105ddad7b68ad0a06d43b21c03cf555dfdba WatchSource:0}: Error finding container 00d5af6e019612533482017f0604105ddad7b68ad0a06d43b21c03cf555dfdba: Status 404 returned error can't find the container with id 00d5af6e019612533482017f0604105ddad7b68ad0a06d43b21c03cf555dfdba Nov 24 17:20:40 crc kubenswrapper[4777]: W1124 17:20:40.594325 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c0b1db4_a8f9_42b1_bb1e_78ee4c5b349b.slice/crio-0209b48607e0f1e400dfd7fbb4e130d91546f003b304134c7ad743793768f0a6 WatchSource:0}: Error finding container 0209b48607e0f1e400dfd7fbb4e130d91546f003b304134c7ad743793768f0a6: Status 404 returned error can't find the container with id 0209b48607e0f1e400dfd7fbb4e130d91546f003b304134c7ad743793768f0a6 Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.620387 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pwl4w" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.635623 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zgxs\" (UniqueName: \"kubernetes.io/projected/f6dac216-d961-42de-9ea6-90ff1126a8ed-kube-api-access-2zgxs\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.635684 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.635706 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdkxf\" (UniqueName: \"kubernetes.io/projected/c2de488a-6bf4-40e2-a403-0f85a3f5177b-kube-api-access-zdkxf\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.635732 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2de488a-6bf4-40e2-a403-0f85a3f5177b-logs\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.635762 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6dac216-d961-42de-9ea6-90ff1126a8ed-logs\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.635777 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.635797 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-config-data\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.635852 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.635873 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-scripts\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.635890 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.635910 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2de488a-6bf4-40e2-a403-0f85a3f5177b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.635949 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.635985 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.636014 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f6dac216-d961-42de-9ea6-90ff1126a8ed-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.659711 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.739644 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-config-data\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.739745 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.739771 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-scripts\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.739819 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.739844 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2de488a-6bf4-40e2-a403-0f85a3f5177b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.739901 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.739930 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.739989 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f6dac216-d961-42de-9ea6-90ff1126a8ed-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.740043 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zgxs\" (UniqueName: \"kubernetes.io/projected/f6dac216-d961-42de-9ea6-90ff1126a8ed-kube-api-access-2zgxs\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.740084 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.740114 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdkxf\" (UniqueName: \"kubernetes.io/projected/c2de488a-6bf4-40e2-a403-0f85a3f5177b-kube-api-access-zdkxf\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.740144 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2de488a-6bf4-40e2-a403-0f85a3f5177b-logs\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.740183 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6dac216-d961-42de-9ea6-90ff1126a8ed-logs\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.740211 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.745040 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2de488a-6bf4-40e2-a403-0f85a3f5177b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.745349 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2de488a-6bf4-40e2-a403-0f85a3f5177b-logs\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.746209 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6dac216-d961-42de-9ea6-90ff1126a8ed-logs\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.746670 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f6dac216-d961-42de-9ea6-90ff1126a8ed-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.748805 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.749055 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0d3411b018636661d0f6478efc7fafc27f5aef76a9aca82320677e283d3b0a36/globalmount\"" pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.752517 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.757589 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-scripts\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.759115 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-config-data\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.759269 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.764228 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.770628 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.784446 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdkxf\" (UniqueName: \"kubernetes.io/projected/c2de488a-6bf4-40e2-a403-0f85a3f5177b-kube-api-access-zdkxf\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.784730 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zgxs\" (UniqueName: \"kubernetes.io/projected/f6dac216-d961-42de-9ea6-90ff1126a8ed-kube-api-access-2zgxs\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.784765 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.784797 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/23594f9d6cc83f7e0c53346caabd46c775bdd206833b4ed7a43e80540ba27612/globalmount\"" pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.934263 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") pod \"glance-default-external-api-0\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.940931 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") pod \"glance-default-internal-api-0\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:20:40 crc kubenswrapper[4777]: I1124 17:20:40.959499 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-57hm5"] Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.001399 4777 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.011804 4777 generic.go:334] "Generic (PLEG): container finished" podID="3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df" containerID="2e53c636590f0d0ff2fd598e95add3c8e39c2a0698bb3d1ce145cd24f81e2a2d" exitCode=0 Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.011859 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" event={"ID":"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df","Type":"ContainerDied","Data":"2e53c636590f0d0ff2fd598e95add3c8e39c2a0698bb3d1ce145cd24f81e2a2d"} Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.034991 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" event={"ID":"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b","Type":"ContainerStarted","Data":"0209b48607e0f1e400dfd7fbb4e130d91546f003b304134c7ad743793768f0a6"} Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.052680 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-57hm5" event={"ID":"a5ad0d8c-d098-4a37-a8f3-2563fff8b620","Type":"ContainerStarted","Data":"7f36932b14ded8561e413452fd2480b8d210336d357414795bb651cc7d9657f4"} Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.054351 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s2227" event={"ID":"62219b70-7aac-4e82-9296-371a2283b6b0","Type":"ContainerStarted","Data":"00d5af6e019612533482017f0604105ddad7b68ad0a06d43b21c03cf555dfdba"} Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.172219 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.186759 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.263375 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25f20c17-1340-4d4e-897a-8ca3230d44b8" path="/var/lib/kubelet/pods/25f20c17-1340-4d4e-897a-8ca3230d44b8/volumes" Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.365817 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-v7hr7"] Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.424278 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.436342 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kl4mk"] Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.664253 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.697672 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-pwl4w"] Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.719535 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-6wwt4"] Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.767359 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-84jqq"] Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.769298 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-dns-svc\") pod \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.770111 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-config\") pod \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.770150 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-ovsdbserver-sb\") pod \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.770180 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-dns-swift-storage-0\") pod \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.770220 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-ovsdbserver-nb\") pod \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.770309 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99kn7\" (UniqueName: \"kubernetes.io/projected/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-kube-api-access-99kn7\") pod \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\" (UID: \"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df\") " Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.785556 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-kube-api-access-99kn7" (OuterVolumeSpecName: "kube-api-access-99kn7") pod "3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df" (UID: "3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df"). InnerVolumeSpecName "kube-api-access-99kn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.833208 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df" (UID: "3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.850025 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df" (UID: "3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.858090 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-config" (OuterVolumeSpecName: "config") pod "3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df" (UID: "3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.862066 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df" (UID: "3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.866589 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df" (UID: "3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.883233 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.883289 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99kn7\" (UniqueName: \"kubernetes.io/projected/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-kube-api-access-99kn7\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.883304 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.883313 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.883323 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:41 crc kubenswrapper[4777]: I1124 17:20:41.883333 4777 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.068015 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.101893 4777 generic.go:334] "Generic (PLEG): container finished" podID="4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b" containerID="ae449764f5cc5b58d4e44f3e140fa3c1f44243d19e457766c1f7ab785c7ebaf7" exitCode=0 Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.102141 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" event={"ID":"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b","Type":"ContainerDied","Data":"ae449764f5cc5b58d4e44f3e140fa3c1f44243d19e457766c1f7ab785c7ebaf7"} Nov 24 17:20:42 crc kubenswrapper[4777]: W1124 17:20:42.114006 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2de488a_6bf4_40e2_a403_0f85a3f5177b.slice/crio-d4b988f9117a869d38f269a4ee00e43a0f7127fb5f66036fe72884eb5698d35a WatchSource:0}: Error finding container d4b988f9117a869d38f269a4ee00e43a0f7127fb5f66036fe72884eb5698d35a: Status 404 returned error can't find the container with id d4b988f9117a869d38f269a4ee00e43a0f7127fb5f66036fe72884eb5698d35a Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.121783 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kl4mk" event={"ID":"b01f3e72-974a-4649-84a3-d23609d774a9","Type":"ContainerStarted","Data":"935312b5eed46e207be26dfa6e5cc349dffef7c3e3093aa5bf5591c49fdaa41b"} Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.121836 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kl4mk" event={"ID":"b01f3e72-974a-4649-84a3-d23609d774a9","Type":"ContainerStarted","Data":"bfec8407a77a8a3403892f589135343ad44c3830e737fe4e4709879774ffc3fa"} Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.134365 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-v7hr7" event={"ID":"d4151db5-e476-4ffa-a038-369874b439cc","Type":"ContainerStarted","Data":"f309353a1a7183b0c92b7a6290f177bbd2df9f7096e08d70d4c6df4f7392333e"} Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.140095 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pwl4w" event={"ID":"d84e561b-fba7-4438-9959-18e1b4bd889c","Type":"ContainerStarted","Data":"6aa2f3918b726f9c1286241aa1f0cce42f1e0d7e5a6945a0ae798b2b76b59ecc"} Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.142049 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s2227" event={"ID":"62219b70-7aac-4e82-9296-371a2283b6b0","Type":"ContainerStarted","Data":"ce26c2638bc8a5a44ac408f9d18326426457cc38ade16ed34b06be16814f2ac4"} Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.146190 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" event={"ID":"3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df","Type":"ContainerDied","Data":"249ebbfc1b122cd30d90b79ba81aa9899a32526f5e730672e8cda507429ad4d2"} Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.146280 4777 scope.go:117] "RemoveContainer" containerID="2e53c636590f0d0ff2fd598e95add3c8e39c2a0698bb3d1ce145cd24f81e2a2d" Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.146387 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-c78nd" Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.147579 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-kl4mk" podStartSLOduration=3.147560454 podStartE2EDuration="3.147560454s" podCreationTimestamp="2025-11-24 17:20:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:20:42.145826155 +0000 UTC m=+1200.305021204" watchObservedRunningTime="2025-11-24 17:20:42.147560454 +0000 UTC m=+1200.306755503" Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.155765 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0eeb390-6ffc-41c8-a816-0b4454730e07","Type":"ContainerStarted","Data":"8341686b30fa5d5d17576f224281e1d58883e70b693778743130240748380fbc"} Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.197378 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6wwt4" event={"ID":"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba","Type":"ContainerStarted","Data":"27aacf67417c552f473245bc1dd5bfe51f1aaa7fff068370bdcec5f149d5b306"} Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.203757 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-s2227" podStartSLOduration=3.20373928 podStartE2EDuration="3.20373928s" podCreationTimestamp="2025-11-24 17:20:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:20:42.196947034 +0000 UTC m=+1200.356142083" watchObservedRunningTime="2025-11-24 17:20:42.20373928 +0000 UTC m=+1200.362934329" Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.219054 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" event={"ID":"14a0dfe7-ab97-4cf8-bd83-542d1410e61d","Type":"ContainerStarted","Data":"2f9f67ae8ec25dabd7c2cb789f85709e201e859b802194400f4d1b48bc74d8cd"} Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.303132 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-c78nd"] Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.311714 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-c78nd"] Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.494324 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.654647 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.882220 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:20:42 crc kubenswrapper[4777]: I1124 17:20:42.968785 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:20:42 crc kubenswrapper[4777]: W1124 17:20:42.974671 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6dac216_d961_42de_9ea6_90ff1126a8ed.slice/crio-8feee448ab7f0e2623fb93cd5e0aed787782cea3bb6e8bbd9a45f5edc9c11311 WatchSource:0}: Error finding container 8feee448ab7f0e2623fb93cd5e0aed787782cea3bb6e8bbd9a45f5edc9c11311: Status 404 returned error can't find the container with id 8feee448ab7f0e2623fb93cd5e0aed787782cea3bb6e8bbd9a45f5edc9c11311 Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.067333 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.122437 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-config\") pod \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.122483 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-ovsdbserver-sb\") pod \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.122512 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-dns-svc\") pod \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.122541 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-ovsdbserver-nb\") pod \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.122666 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwl56\" (UniqueName: \"kubernetes.io/projected/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-kube-api-access-gwl56\") pod \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.122822 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-dns-swift-storage-0\") pod \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\" (UID: \"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b\") " Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.144082 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-kube-api-access-gwl56" (OuterVolumeSpecName: "kube-api-access-gwl56") pod "4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b" (UID: "4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b"). InnerVolumeSpecName "kube-api-access-gwl56". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.187109 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-config" (OuterVolumeSpecName: "config") pod "4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b" (UID: "4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.210505 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b" (UID: "4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.212411 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b" (UID: "4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.213318 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b" (UID: "4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.237902 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b" (UID: "4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.245089 4777 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.245117 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.245127 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.245136 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.245145 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.245160 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwl56\" (UniqueName: \"kubernetes.io/projected/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b-kube-api-access-gwl56\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.253815 4777 generic.go:334] "Generic (PLEG): container finished" podID="14a0dfe7-ab97-4cf8-bd83-542d1410e61d" containerID="638b6c66ff3b9de454777656942a7865c7adcf89f0fcdfe31f018ba89175be80" exitCode=0 Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.260805 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.316035 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df" path="/var/lib/kubelet/pods/3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df/volumes" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.317533 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-cxtnv" event={"ID":"4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b","Type":"ContainerDied","Data":"0209b48607e0f1e400dfd7fbb4e130d91546f003b304134c7ad743793768f0a6"} Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.317587 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" event={"ID":"14a0dfe7-ab97-4cf8-bd83-542d1410e61d","Type":"ContainerDied","Data":"638b6c66ff3b9de454777656942a7865c7adcf89f0fcdfe31f018ba89175be80"} Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.317604 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f6dac216-d961-42de-9ea6-90ff1126a8ed","Type":"ContainerStarted","Data":"8feee448ab7f0e2623fb93cd5e0aed787782cea3bb6e8bbd9a45f5edc9c11311"} Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.317616 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c2de488a-6bf4-40e2-a403-0f85a3f5177b","Type":"ContainerStarted","Data":"d4b988f9117a869d38f269a4ee00e43a0f7127fb5f66036fe72884eb5698d35a"} Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.317637 4777 scope.go:117] "RemoveContainer" containerID="ae449764f5cc5b58d4e44f3e140fa3c1f44243d19e457766c1f7ab785c7ebaf7" Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.490478 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-cxtnv"] Nov 24 17:20:43 crc kubenswrapper[4777]: I1124 17:20:43.517605 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-cxtnv"] Nov 24 17:20:44 crc kubenswrapper[4777]: I1124 17:20:44.308041 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" event={"ID":"14a0dfe7-ab97-4cf8-bd83-542d1410e61d","Type":"ContainerStarted","Data":"ca273ad75ed251795bf946baba0feb47292d34f34d6aba8b1d4f27ebc5f8b270"} Nov 24 17:20:44 crc kubenswrapper[4777]: I1124 17:20:44.308556 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:44 crc kubenswrapper[4777]: I1124 17:20:44.315050 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c2de488a-6bf4-40e2-a403-0f85a3f5177b","Type":"ContainerStarted","Data":"7916992607b47ed275695dd63f03f23afa8871f7286c92ff72e8318e74866b0f"} Nov 24 17:20:44 crc kubenswrapper[4777]: I1124 17:20:44.328304 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" podStartSLOduration=5.328282497 podStartE2EDuration="5.328282497s" podCreationTimestamp="2025-11-24 17:20:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:20:44.325752114 +0000 UTC m=+1202.484947163" watchObservedRunningTime="2025-11-24 17:20:44.328282497 +0000 UTC m=+1202.487477546" Nov 24 17:20:45 crc kubenswrapper[4777]: I1124 17:20:45.260105 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b" path="/var/lib/kubelet/pods/4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b/volumes" Nov 24 17:20:45 crc kubenswrapper[4777]: I1124 17:20:45.332013 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c2de488a-6bf4-40e2-a403-0f85a3f5177b","Type":"ContainerStarted","Data":"29523a78f037d3852a1e479c36e9f0e287f315efb31eadb341215455516829f7"} Nov 24 17:20:45 crc kubenswrapper[4777]: I1124 17:20:45.332213 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c2de488a-6bf4-40e2-a403-0f85a3f5177b" containerName="glance-log" containerID="cri-o://7916992607b47ed275695dd63f03f23afa8871f7286c92ff72e8318e74866b0f" gracePeriod=30 Nov 24 17:20:45 crc kubenswrapper[4777]: I1124 17:20:45.332377 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c2de488a-6bf4-40e2-a403-0f85a3f5177b" containerName="glance-httpd" containerID="cri-o://29523a78f037d3852a1e479c36e9f0e287f315efb31eadb341215455516829f7" gracePeriod=30 Nov 24 17:20:45 crc kubenswrapper[4777]: I1124 17:20:45.337910 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f6dac216-d961-42de-9ea6-90ff1126a8ed" containerName="glance-log" containerID="cri-o://5a67315ae43c2a6697812e5a7b6f9cc93c9ed13f3a104aeb98979a1cf9072dde" gracePeriod=30 Nov 24 17:20:45 crc kubenswrapper[4777]: I1124 17:20:45.338043 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f6dac216-d961-42de-9ea6-90ff1126a8ed" containerName="glance-httpd" containerID="cri-o://1740992d52d331d743216a96923b96754484801652f49b7638814c53dc620811" gracePeriod=30 Nov 24 17:20:45 crc kubenswrapper[4777]: I1124 17:20:45.338095 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f6dac216-d961-42de-9ea6-90ff1126a8ed","Type":"ContainerStarted","Data":"1740992d52d331d743216a96923b96754484801652f49b7638814c53dc620811"} Nov 24 17:20:45 crc kubenswrapper[4777]: I1124 17:20:45.338116 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f6dac216-d961-42de-9ea6-90ff1126a8ed","Type":"ContainerStarted","Data":"5a67315ae43c2a6697812e5a7b6f9cc93c9ed13f3a104aeb98979a1cf9072dde"} Nov 24 17:20:45 crc kubenswrapper[4777]: I1124 17:20:45.362565 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.362541924 podStartE2EDuration="6.362541924s" podCreationTimestamp="2025-11-24 17:20:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:20:45.351653661 +0000 UTC m=+1203.510848710" watchObservedRunningTime="2025-11-24 17:20:45.362541924 +0000 UTC m=+1203.521736973" Nov 24 17:20:46 crc kubenswrapper[4777]: I1124 17:20:46.357292 4777 generic.go:334] "Generic (PLEG): container finished" podID="c2de488a-6bf4-40e2-a403-0f85a3f5177b" containerID="29523a78f037d3852a1e479c36e9f0e287f315efb31eadb341215455516829f7" exitCode=0 Nov 24 17:20:46 crc kubenswrapper[4777]: I1124 17:20:46.361042 4777 generic.go:334] "Generic (PLEG): container finished" podID="c2de488a-6bf4-40e2-a403-0f85a3f5177b" containerID="7916992607b47ed275695dd63f03f23afa8871f7286c92ff72e8318e74866b0f" exitCode=143 Nov 24 17:20:46 crc kubenswrapper[4777]: I1124 17:20:46.357345 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c2de488a-6bf4-40e2-a403-0f85a3f5177b","Type":"ContainerDied","Data":"29523a78f037d3852a1e479c36e9f0e287f315efb31eadb341215455516829f7"} Nov 24 17:20:46 crc kubenswrapper[4777]: I1124 17:20:46.361341 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c2de488a-6bf4-40e2-a403-0f85a3f5177b","Type":"ContainerDied","Data":"7916992607b47ed275695dd63f03f23afa8871f7286c92ff72e8318e74866b0f"} Nov 24 17:20:46 crc kubenswrapper[4777]: I1124 17:20:46.369173 4777 generic.go:334] "Generic (PLEG): container finished" podID="f6dac216-d961-42de-9ea6-90ff1126a8ed" containerID="1740992d52d331d743216a96923b96754484801652f49b7638814c53dc620811" exitCode=143 Nov 24 17:20:46 crc kubenswrapper[4777]: I1124 17:20:46.369199 4777 generic.go:334] "Generic (PLEG): container finished" podID="f6dac216-d961-42de-9ea6-90ff1126a8ed" containerID="5a67315ae43c2a6697812e5a7b6f9cc93c9ed13f3a104aeb98979a1cf9072dde" exitCode=143 Nov 24 17:20:46 crc kubenswrapper[4777]: I1124 17:20:46.369301 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f6dac216-d961-42de-9ea6-90ff1126a8ed","Type":"ContainerDied","Data":"1740992d52d331d743216a96923b96754484801652f49b7638814c53dc620811"} Nov 24 17:20:46 crc kubenswrapper[4777]: I1124 17:20:46.369364 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f6dac216-d961-42de-9ea6-90ff1126a8ed","Type":"ContainerDied","Data":"5a67315ae43c2a6697812e5a7b6f9cc93c9ed13f3a104aeb98979a1cf9072dde"} Nov 24 17:20:46 crc kubenswrapper[4777]: I1124 17:20:46.373158 4777 generic.go:334] "Generic (PLEG): container finished" podID="62219b70-7aac-4e82-9296-371a2283b6b0" containerID="ce26c2638bc8a5a44ac408f9d18326426457cc38ade16ed34b06be16814f2ac4" exitCode=0 Nov 24 17:20:46 crc kubenswrapper[4777]: I1124 17:20:46.373192 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s2227" event={"ID":"62219b70-7aac-4e82-9296-371a2283b6b0","Type":"ContainerDied","Data":"ce26c2638bc8a5a44ac408f9d18326426457cc38ade16ed34b06be16814f2ac4"} Nov 24 17:20:46 crc kubenswrapper[4777]: I1124 17:20:46.389806 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.389787741 podStartE2EDuration="7.389787741s" podCreationTimestamp="2025-11-24 17:20:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:20:45.379214114 +0000 UTC m=+1203.538409163" watchObservedRunningTime="2025-11-24 17:20:46.389787741 +0000 UTC m=+1204.548982790" Nov 24 17:20:50 crc kubenswrapper[4777]: I1124 17:20:50.661270 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:20:50 crc kubenswrapper[4777]: I1124 17:20:50.729520 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-fq2lk"] Nov 24 17:20:50 crc kubenswrapper[4777]: I1124 17:20:50.729748 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" podUID="10e17cd8-dd7d-476a-96f7-27ac1f938b83" containerName="dnsmasq-dns" containerID="cri-o://e0c118edd84e3687eb01c0c0fec0d5fce1f55f30cf6342ad71f957d16cd34ea7" gracePeriod=10 Nov 24 17:20:51 crc kubenswrapper[4777]: I1124 17:20:51.556106 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" podUID="10e17cd8-dd7d-476a-96f7-27ac1f938b83" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: connect: connection refused" Nov 24 17:20:53 crc kubenswrapper[4777]: I1124 17:20:53.452108 4777 generic.go:334] "Generic (PLEG): container finished" podID="10e17cd8-dd7d-476a-96f7-27ac1f938b83" containerID="e0c118edd84e3687eb01c0c0fec0d5fce1f55f30cf6342ad71f957d16cd34ea7" exitCode=0 Nov 24 17:20:53 crc kubenswrapper[4777]: I1124 17:20:53.452244 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" event={"ID":"10e17cd8-dd7d-476a-96f7-27ac1f938b83","Type":"ContainerDied","Data":"e0c118edd84e3687eb01c0c0fec0d5fce1f55f30cf6342ad71f957d16cd34ea7"} Nov 24 17:20:56 crc kubenswrapper[4777]: I1124 17:20:56.555412 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" podUID="10e17cd8-dd7d-476a-96f7-27ac1f938b83" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: connect: connection refused" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.298348 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.381505 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-config-data\") pod \"62219b70-7aac-4e82-9296-371a2283b6b0\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.381581 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-credential-keys\") pod \"62219b70-7aac-4e82-9296-371a2283b6b0\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.381660 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-fernet-keys\") pod \"62219b70-7aac-4e82-9296-371a2283b6b0\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.381727 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l66n8\" (UniqueName: \"kubernetes.io/projected/62219b70-7aac-4e82-9296-371a2283b6b0-kube-api-access-l66n8\") pod \"62219b70-7aac-4e82-9296-371a2283b6b0\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.381807 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-scripts\") pod \"62219b70-7aac-4e82-9296-371a2283b6b0\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.381820 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-combined-ca-bundle\") pod \"62219b70-7aac-4e82-9296-371a2283b6b0\" (UID: \"62219b70-7aac-4e82-9296-371a2283b6b0\") " Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.387858 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "62219b70-7aac-4e82-9296-371a2283b6b0" (UID: "62219b70-7aac-4e82-9296-371a2283b6b0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.388274 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62219b70-7aac-4e82-9296-371a2283b6b0-kube-api-access-l66n8" (OuterVolumeSpecName: "kube-api-access-l66n8") pod "62219b70-7aac-4e82-9296-371a2283b6b0" (UID: "62219b70-7aac-4e82-9296-371a2283b6b0"). InnerVolumeSpecName "kube-api-access-l66n8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.388598 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-scripts" (OuterVolumeSpecName: "scripts") pod "62219b70-7aac-4e82-9296-371a2283b6b0" (UID: "62219b70-7aac-4e82-9296-371a2283b6b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.389187 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "62219b70-7aac-4e82-9296-371a2283b6b0" (UID: "62219b70-7aac-4e82-9296-371a2283b6b0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.412926 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62219b70-7aac-4e82-9296-371a2283b6b0" (UID: "62219b70-7aac-4e82-9296-371a2283b6b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.416201 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-config-data" (OuterVolumeSpecName: "config-data") pod "62219b70-7aac-4e82-9296-371a2283b6b0" (UID: "62219b70-7aac-4e82-9296-371a2283b6b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.484421 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.484462 4777 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.484477 4777 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.484492 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l66n8\" (UniqueName: \"kubernetes.io/projected/62219b70-7aac-4e82-9296-371a2283b6b0-kube-api-access-l66n8\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.484505 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.484515 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62219b70-7aac-4e82-9296-371a2283b6b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.501388 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s2227" event={"ID":"62219b70-7aac-4e82-9296-371a2283b6b0","Type":"ContainerDied","Data":"00d5af6e019612533482017f0604105ddad7b68ad0a06d43b21c03cf555dfdba"} Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.501430 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s2227" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.501429 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00d5af6e019612533482017f0604105ddad7b68ad0a06d43b21c03cf555dfdba" Nov 24 17:20:57 crc kubenswrapper[4777]: E1124 17:20:57.803696 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 24 17:20:57 crc kubenswrapper[4777]: E1124 17:20:57.804822 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rszcl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-v7hr7_openstack(d4151db5-e476-4ffa-a038-369874b439cc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:20:57 crc kubenswrapper[4777]: E1124 17:20:57.806213 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-v7hr7" podUID="d4151db5-e476-4ffa-a038-369874b439cc" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.858007 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.993547 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-config-data\") pod \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.993631 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-combined-ca-bundle\") pod \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.993696 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdkxf\" (UniqueName: \"kubernetes.io/projected/c2de488a-6bf4-40e2-a403-0f85a3f5177b-kube-api-access-zdkxf\") pod \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.993805 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") pod \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.993854 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-scripts\") pod \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.993899 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2de488a-6bf4-40e2-a403-0f85a3f5177b-httpd-run\") pod \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.993925 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2de488a-6bf4-40e2-a403-0f85a3f5177b-logs\") pod \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\" (UID: \"c2de488a-6bf4-40e2-a403-0f85a3f5177b\") " Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.994619 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2de488a-6bf4-40e2-a403-0f85a3f5177b-logs" (OuterVolumeSpecName: "logs") pod "c2de488a-6bf4-40e2-a403-0f85a3f5177b" (UID: "c2de488a-6bf4-40e2-a403-0f85a3f5177b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:20:57 crc kubenswrapper[4777]: I1124 17:20:57.994687 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2de488a-6bf4-40e2-a403-0f85a3f5177b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c2de488a-6bf4-40e2-a403-0f85a3f5177b" (UID: "c2de488a-6bf4-40e2-a403-0f85a3f5177b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.000120 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-scripts" (OuterVolumeSpecName: "scripts") pod "c2de488a-6bf4-40e2-a403-0f85a3f5177b" (UID: "c2de488a-6bf4-40e2-a403-0f85a3f5177b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.000137 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2de488a-6bf4-40e2-a403-0f85a3f5177b-kube-api-access-zdkxf" (OuterVolumeSpecName: "kube-api-access-zdkxf") pod "c2de488a-6bf4-40e2-a403-0f85a3f5177b" (UID: "c2de488a-6bf4-40e2-a403-0f85a3f5177b"). InnerVolumeSpecName "kube-api-access-zdkxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.007963 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8" (OuterVolumeSpecName: "glance") pod "c2de488a-6bf4-40e2-a403-0f85a3f5177b" (UID: "c2de488a-6bf4-40e2-a403-0f85a3f5177b"). InnerVolumeSpecName "pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.058303 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2de488a-6bf4-40e2-a403-0f85a3f5177b" (UID: "c2de488a-6bf4-40e2-a403-0f85a3f5177b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.062303 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-config-data" (OuterVolumeSpecName: "config-data") pod "c2de488a-6bf4-40e2-a403-0f85a3f5177b" (UID: "c2de488a-6bf4-40e2-a403-0f85a3f5177b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.095715 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdkxf\" (UniqueName: \"kubernetes.io/projected/c2de488a-6bf4-40e2-a403-0f85a3f5177b-kube-api-access-zdkxf\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.095769 4777 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") on node \"crc\" " Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.095781 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.095792 4777 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2de488a-6bf4-40e2-a403-0f85a3f5177b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.095804 4777 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2de488a-6bf4-40e2-a403-0f85a3f5177b-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.095813 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.095821 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2de488a-6bf4-40e2-a403-0f85a3f5177b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.121926 4777 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.122082 4777 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8") on node "crc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.197584 4777 reconciler_common.go:293] "Volume detached for volume \"pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") on node \"crc\" DevicePath \"\"" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.442471 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-s2227"] Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.452729 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-s2227"] Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.516858 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.517027 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c2de488a-6bf4-40e2-a403-0f85a3f5177b","Type":"ContainerDied","Data":"d4b988f9117a869d38f269a4ee00e43a0f7127fb5f66036fe72884eb5698d35a"} Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.517081 4777 scope.go:117] "RemoveContainer" containerID="29523a78f037d3852a1e479c36e9f0e287f315efb31eadb341215455516829f7" Nov 24 17:20:58 crc kubenswrapper[4777]: E1124 17:20:58.524438 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-v7hr7" podUID="d4151db5-e476-4ffa-a038-369874b439cc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.542557 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-k48pc"] Nov 24 17:20:58 crc kubenswrapper[4777]: E1124 17:20:58.544432 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df" containerName="init" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.544619 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df" containerName="init" Nov 24 17:20:58 crc kubenswrapper[4777]: E1124 17:20:58.544642 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b" containerName="init" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.544650 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b" containerName="init" Nov 24 17:20:58 crc kubenswrapper[4777]: E1124 17:20:58.544670 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2de488a-6bf4-40e2-a403-0f85a3f5177b" containerName="glance-httpd" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.544677 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2de488a-6bf4-40e2-a403-0f85a3f5177b" containerName="glance-httpd" Nov 24 17:20:58 crc kubenswrapper[4777]: E1124 17:20:58.544694 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62219b70-7aac-4e82-9296-371a2283b6b0" containerName="keystone-bootstrap" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.544699 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="62219b70-7aac-4e82-9296-371a2283b6b0" containerName="keystone-bootstrap" Nov 24 17:20:58 crc kubenswrapper[4777]: E1124 17:20:58.544710 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2de488a-6bf4-40e2-a403-0f85a3f5177b" containerName="glance-log" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.544715 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2de488a-6bf4-40e2-a403-0f85a3f5177b" containerName="glance-log" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.544922 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e5d25a6-4e3f-4b0a-8b0e-bee6c2f2f5df" containerName="init" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.544940 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c0b1db4-a8f9-42b1-bb1e-78ee4c5b349b" containerName="init" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.544952 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="62219b70-7aac-4e82-9296-371a2283b6b0" containerName="keystone-bootstrap" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.544959 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2de488a-6bf4-40e2-a403-0f85a3f5177b" containerName="glance-log" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.544986 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2de488a-6bf4-40e2-a403-0f85a3f5177b" containerName="glance-httpd" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.545750 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.552210 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.552602 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.552822 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vzrxj" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.553123 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.563853 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-k48pc"] Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.599924 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.605422 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-combined-ca-bundle\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.605621 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-config-data\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.605710 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-scripts\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.605848 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp6rp\" (UniqueName: \"kubernetes.io/projected/6d2fef26-a53d-41dd-8523-b451f2ea8127-kube-api-access-dp6rp\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.606047 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-fernet-keys\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.606144 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-credential-keys\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.609056 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.618015 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.620262 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.623405 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.623428 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.626320 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.708471 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp6rp\" (UniqueName: \"kubernetes.io/projected/6d2fef26-a53d-41dd-8523-b451f2ea8127-kube-api-access-dp6rp\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.708573 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-scripts\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.708608 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.708651 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-fernet-keys\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.708808 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-credential-keys\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.708879 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szvz2\" (UniqueName: \"kubernetes.io/projected/5b3890d7-b149-4be6-bff8-60f8b7de9e22-kube-api-access-szvz2\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.708931 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b3890d7-b149-4be6-bff8-60f8b7de9e22-logs\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.709040 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5b3890d7-b149-4be6-bff8-60f8b7de9e22-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.709070 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-combined-ca-bundle\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.709112 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.709218 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-config-data\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.709242 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-config-data\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.709289 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-scripts\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.709339 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.712633 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-combined-ca-bundle\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.713012 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-scripts\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.713221 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-credential-keys\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.724247 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp6rp\" (UniqueName: \"kubernetes.io/projected/6d2fef26-a53d-41dd-8523-b451f2ea8127-kube-api-access-dp6rp\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.729165 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-fernet-keys\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.731122 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-config-data\") pod \"keystone-bootstrap-k48pc\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.810927 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-scripts\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.811007 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.811081 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szvz2\" (UniqueName: \"kubernetes.io/projected/5b3890d7-b149-4be6-bff8-60f8b7de9e22-kube-api-access-szvz2\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.811106 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b3890d7-b149-4be6-bff8-60f8b7de9e22-logs\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.811125 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5b3890d7-b149-4be6-bff8-60f8b7de9e22-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.811152 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.811193 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-config-data\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.811225 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.812219 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b3890d7-b149-4be6-bff8-60f8b7de9e22-logs\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.812483 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5b3890d7-b149-4be6-bff8-60f8b7de9e22-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.815321 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.815659 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-scripts\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.816570 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-config-data\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.816590 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.816699 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/23594f9d6cc83f7e0c53346caabd46c775bdd206833b4ed7a43e80540ba27612/globalmount\"" pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.820500 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.832437 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szvz2\" (UniqueName: \"kubernetes.io/projected/5b3890d7-b149-4be6-bff8-60f8b7de9e22-kube-api-access-szvz2\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.850997 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") pod \"glance-default-external-api-0\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " pod="openstack/glance-default-external-api-0" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.873548 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:20:58 crc kubenswrapper[4777]: I1124 17:20:58.951057 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:20:59 crc kubenswrapper[4777]: I1124 17:20:59.260910 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62219b70-7aac-4e82-9296-371a2283b6b0" path="/var/lib/kubelet/pods/62219b70-7aac-4e82-9296-371a2283b6b0/volumes" Nov 24 17:20:59 crc kubenswrapper[4777]: I1124 17:20:59.261908 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2de488a-6bf4-40e2-a403-0f85a3f5177b" path="/var/lib/kubelet/pods/c2de488a-6bf4-40e2-a403-0f85a3f5177b/volumes" Nov 24 17:21:06 crc kubenswrapper[4777]: I1124 17:21:06.554813 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" podUID="10e17cd8-dd7d-476a-96f7-27ac1f938b83" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: i/o timeout" Nov 24 17:21:06 crc kubenswrapper[4777]: I1124 17:21:06.555785 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:21:11 crc kubenswrapper[4777]: I1124 17:21:11.173517 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 17:21:11 crc kubenswrapper[4777]: I1124 17:21:11.174110 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 17:21:11 crc kubenswrapper[4777]: I1124 17:21:11.556554 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" podUID="10e17cd8-dd7d-476a-96f7-27ac1f938b83" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: i/o timeout" Nov 24 17:21:13 crc kubenswrapper[4777]: E1124 17:21:13.410359 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Nov 24 17:21:13 crc kubenswrapper[4777]: E1124 17:21:13.411004 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n66fh5cdh67ch68fh57bh667h85h67ch66dh599h5b5h64dh65fhch87hfh68chb4h5b8h667h6fh59bh5f4h54ch549hbfh575h88h5b7hb4h54dh547q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qd7sx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(c0eeb390-6ffc-41c8-a816-0b4454730e07): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.457794 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.464653 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.628997 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-combined-ca-bundle\") pod \"f6dac216-d961-42de-9ea6-90ff1126a8ed\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.629063 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-ovsdbserver-sb\") pod \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.629141 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zgxs\" (UniqueName: \"kubernetes.io/projected/f6dac216-d961-42de-9ea6-90ff1126a8ed-kube-api-access-2zgxs\") pod \"f6dac216-d961-42de-9ea6-90ff1126a8ed\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.629169 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-ovsdbserver-nb\") pod \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.629193 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42r8s\" (UniqueName: \"kubernetes.io/projected/10e17cd8-dd7d-476a-96f7-27ac1f938b83-kube-api-access-42r8s\") pod \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.629246 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f6dac216-d961-42de-9ea6-90ff1126a8ed-httpd-run\") pod \"f6dac216-d961-42de-9ea6-90ff1126a8ed\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.629270 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-config\") pod \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.629482 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") pod \"f6dac216-d961-42de-9ea6-90ff1126a8ed\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.629513 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-config-data\") pod \"f6dac216-d961-42de-9ea6-90ff1126a8ed\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.629554 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-scripts\") pod \"f6dac216-d961-42de-9ea6-90ff1126a8ed\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.629620 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6dac216-d961-42de-9ea6-90ff1126a8ed-logs\") pod \"f6dac216-d961-42de-9ea6-90ff1126a8ed\" (UID: \"f6dac216-d961-42de-9ea6-90ff1126a8ed\") " Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.629638 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-dns-svc\") pod \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\" (UID: \"10e17cd8-dd7d-476a-96f7-27ac1f938b83\") " Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.629874 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6dac216-d961-42de-9ea6-90ff1126a8ed-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f6dac216-d961-42de-9ea6-90ff1126a8ed" (UID: "f6dac216-d961-42de-9ea6-90ff1126a8ed"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.630218 4777 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f6dac216-d961-42de-9ea6-90ff1126a8ed-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.633328 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6dac216-d961-42de-9ea6-90ff1126a8ed-logs" (OuterVolumeSpecName: "logs") pod "f6dac216-d961-42de-9ea6-90ff1126a8ed" (UID: "f6dac216-d961-42de-9ea6-90ff1126a8ed"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.635938 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10e17cd8-dd7d-476a-96f7-27ac1f938b83-kube-api-access-42r8s" (OuterVolumeSpecName: "kube-api-access-42r8s") pod "10e17cd8-dd7d-476a-96f7-27ac1f938b83" (UID: "10e17cd8-dd7d-476a-96f7-27ac1f938b83"). InnerVolumeSpecName "kube-api-access-42r8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.641174 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6dac216-d961-42de-9ea6-90ff1126a8ed-kube-api-access-2zgxs" (OuterVolumeSpecName: "kube-api-access-2zgxs") pod "f6dac216-d961-42de-9ea6-90ff1126a8ed" (UID: "f6dac216-d961-42de-9ea6-90ff1126a8ed"). InnerVolumeSpecName "kube-api-access-2zgxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.644560 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-scripts" (OuterVolumeSpecName: "scripts") pod "f6dac216-d961-42de-9ea6-90ff1126a8ed" (UID: "f6dac216-d961-42de-9ea6-90ff1126a8ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.646589 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4" (OuterVolumeSpecName: "glance") pod "f6dac216-d961-42de-9ea6-90ff1126a8ed" (UID: "f6dac216-d961-42de-9ea6-90ff1126a8ed"). InnerVolumeSpecName "pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.671468 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6dac216-d961-42de-9ea6-90ff1126a8ed" (UID: "f6dac216-d961-42de-9ea6-90ff1126a8ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.679142 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" event={"ID":"10e17cd8-dd7d-476a-96f7-27ac1f938b83","Type":"ContainerDied","Data":"4648cae05bb4802db0f8d420331a95ab758759c73ef7b24a03a2d81dfdce3e88"} Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.679244 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.686208 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f6dac216-d961-42de-9ea6-90ff1126a8ed","Type":"ContainerDied","Data":"8feee448ab7f0e2623fb93cd5e0aed787782cea3bb6e8bbd9a45f5edc9c11311"} Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.686316 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.694131 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "10e17cd8-dd7d-476a-96f7-27ac1f938b83" (UID: "10e17cd8-dd7d-476a-96f7-27ac1f938b83"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.703162 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-config-data" (OuterVolumeSpecName: "config-data") pod "f6dac216-d961-42de-9ea6-90ff1126a8ed" (UID: "f6dac216-d961-42de-9ea6-90ff1126a8ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.708194 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "10e17cd8-dd7d-476a-96f7-27ac1f938b83" (UID: "10e17cd8-dd7d-476a-96f7-27ac1f938b83"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.709279 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "10e17cd8-dd7d-476a-96f7-27ac1f938b83" (UID: "10e17cd8-dd7d-476a-96f7-27ac1f938b83"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.717304 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-config" (OuterVolumeSpecName: "config") pod "10e17cd8-dd7d-476a-96f7-27ac1f938b83" (UID: "10e17cd8-dd7d-476a-96f7-27ac1f938b83"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.732067 4777 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") on node \"crc\" " Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.732103 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.732115 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.732124 4777 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6dac216-d961-42de-9ea6-90ff1126a8ed-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.732134 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.732144 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6dac216-d961-42de-9ea6-90ff1126a8ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.732157 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.732165 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zgxs\" (UniqueName: \"kubernetes.io/projected/f6dac216-d961-42de-9ea6-90ff1126a8ed-kube-api-access-2zgxs\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.732173 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.732182 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42r8s\" (UniqueName: \"kubernetes.io/projected/10e17cd8-dd7d-476a-96f7-27ac1f938b83-kube-api-access-42r8s\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.732190 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e17cd8-dd7d-476a-96f7-27ac1f938b83-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.754739 4777 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.754873 4777 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4") on node "crc" Nov 24 17:21:13 crc kubenswrapper[4777]: I1124 17:21:13.834238 4777 reconciler_common.go:293] "Volume detached for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.027469 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-fq2lk"] Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.042352 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-fq2lk"] Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.058959 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.077925 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.089189 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:21:14 crc kubenswrapper[4777]: E1124 17:21:14.089576 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10e17cd8-dd7d-476a-96f7-27ac1f938b83" containerName="init" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.089590 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="10e17cd8-dd7d-476a-96f7-27ac1f938b83" containerName="init" Nov 24 17:21:14 crc kubenswrapper[4777]: E1124 17:21:14.089605 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6dac216-d961-42de-9ea6-90ff1126a8ed" containerName="glance-httpd" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.089612 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6dac216-d961-42de-9ea6-90ff1126a8ed" containerName="glance-httpd" Nov 24 17:21:14 crc kubenswrapper[4777]: E1124 17:21:14.089634 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6dac216-d961-42de-9ea6-90ff1126a8ed" containerName="glance-log" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.089643 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6dac216-d961-42de-9ea6-90ff1126a8ed" containerName="glance-log" Nov 24 17:21:14 crc kubenswrapper[4777]: E1124 17:21:14.089654 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10e17cd8-dd7d-476a-96f7-27ac1f938b83" containerName="dnsmasq-dns" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.089661 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="10e17cd8-dd7d-476a-96f7-27ac1f938b83" containerName="dnsmasq-dns" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.089875 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6dac216-d961-42de-9ea6-90ff1126a8ed" containerName="glance-httpd" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.089898 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6dac216-d961-42de-9ea6-90ff1126a8ed" containerName="glance-log" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.089911 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="10e17cd8-dd7d-476a-96f7-27ac1f938b83" containerName="dnsmasq-dns" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.092405 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.095526 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.095754 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.096053 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.243609 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.243657 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.243683 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d03c2407-e0ad-49da-91a9-c9d3e8e67692-logs\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.243749 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m95pf\" (UniqueName: \"kubernetes.io/projected/d03c2407-e0ad-49da-91a9-c9d3e8e67692-kube-api-access-m95pf\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.243917 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.243985 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.244070 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d03c2407-e0ad-49da-91a9-c9d3e8e67692-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.244141 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.345930 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.346010 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d03c2407-e0ad-49da-91a9-c9d3e8e67692-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.346047 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.346102 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.346123 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.346146 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d03c2407-e0ad-49da-91a9-c9d3e8e67692-logs\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.346204 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m95pf\" (UniqueName: \"kubernetes.io/projected/d03c2407-e0ad-49da-91a9-c9d3e8e67692-kube-api-access-m95pf\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.346246 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.347774 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d03c2407-e0ad-49da-91a9-c9d3e8e67692-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.347982 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d03c2407-e0ad-49da-91a9-c9d3e8e67692-logs\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.351400 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.352752 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.352784 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0d3411b018636661d0f6478efc7fafc27f5aef76a9aca82320677e283d3b0a36/globalmount\"" pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.355783 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.359862 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.360455 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.369645 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m95pf\" (UniqueName: \"kubernetes.io/projected/d03c2407-e0ad-49da-91a9-c9d3e8e67692-kube-api-access-m95pf\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.393128 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") pod \"glance-default-internal-api-0\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:21:14 crc kubenswrapper[4777]: I1124 17:21:14.408841 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:21:15 crc kubenswrapper[4777]: I1124 17:21:15.258847 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10e17cd8-dd7d-476a-96f7-27ac1f938b83" path="/var/lib/kubelet/pods/10e17cd8-dd7d-476a-96f7-27ac1f938b83/volumes" Nov 24 17:21:15 crc kubenswrapper[4777]: I1124 17:21:15.260512 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6dac216-d961-42de-9ea6-90ff1126a8ed" path="/var/lib/kubelet/pods/f6dac216-d961-42de-9ea6-90ff1126a8ed/volumes" Nov 24 17:21:15 crc kubenswrapper[4777]: I1124 17:21:15.271761 4777 scope.go:117] "RemoveContainer" containerID="7916992607b47ed275695dd63f03f23afa8871f7286c92ff72e8318e74866b0f" Nov 24 17:21:15 crc kubenswrapper[4777]: E1124 17:21:15.289296 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 24 17:21:15 crc kubenswrapper[4777]: E1124 17:21:15.289503 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5l8gn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-6wwt4_openstack(95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:21:15 crc kubenswrapper[4777]: E1124 17:21:15.290733 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-6wwt4" podUID="95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba" Nov 24 17:21:15 crc kubenswrapper[4777]: E1124 17:21:15.710732 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-6wwt4" podUID="95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba" Nov 24 17:21:16 crc kubenswrapper[4777]: I1124 17:21:16.556953 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-fq2lk" podUID="10e17cd8-dd7d-476a-96f7-27ac1f938b83" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: i/o timeout" Nov 24 17:21:19 crc kubenswrapper[4777]: I1124 17:21:19.471864 4777 scope.go:117] "RemoveContainer" containerID="e0c118edd84e3687eb01c0c0fec0d5fce1f55f30cf6342ad71f957d16cd34ea7" Nov 24 17:21:19 crc kubenswrapper[4777]: I1124 17:21:19.882483 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-k48pc"] Nov 24 17:21:20 crc kubenswrapper[4777]: I1124 17:21:20.207019 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:21:20 crc kubenswrapper[4777]: E1124 17:21:20.393144 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current" Nov 24 17:21:20 crc kubenswrapper[4777]: E1124 17:21:20.393492 4777 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current" Nov 24 17:21:20 crc kubenswrapper[4777]: E1124 17:21:20.393648 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cloudkitty-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CloudKittyPassword,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:CloudKittyPassword,Optional:nil,},},},EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:cloudkitty-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:certs,ReadOnly:true,MountPath:/var/lib/openstack/loki-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zszjk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42406,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-db-sync-57hm5_openstack(a5ad0d8c-d098-4a37-a8f3-2563fff8b620): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:21:20 crc kubenswrapper[4777]: E1124 17:21:20.394864 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cloudkitty-db-sync-57hm5" podUID="a5ad0d8c-d098-4a37-a8f3-2563fff8b620" Nov 24 17:21:20 crc kubenswrapper[4777]: W1124 17:21:20.689935 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d2fef26_a53d_41dd_8523_b451f2ea8127.slice/crio-772aac6beaca59b436aba0c8bc944aad346788833a5922ff2ee7deaecb1cea4c WatchSource:0}: Error finding container 772aac6beaca59b436aba0c8bc944aad346788833a5922ff2ee7deaecb1cea4c: Status 404 returned error can't find the container with id 772aac6beaca59b436aba0c8bc944aad346788833a5922ff2ee7deaecb1cea4c Nov 24 17:21:20 crc kubenswrapper[4777]: W1124 17:21:20.698373 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b3890d7_b149_4be6_bff8_60f8b7de9e22.slice/crio-72433a3f6fdb73cecb9f1ff2e4d32735952d2974d6c98d8033a7821ff9d83cc1 WatchSource:0}: Error finding container 72433a3f6fdb73cecb9f1ff2e4d32735952d2974d6c98d8033a7821ff9d83cc1: Status 404 returned error can't find the container with id 72433a3f6fdb73cecb9f1ff2e4d32735952d2974d6c98d8033a7821ff9d83cc1 Nov 24 17:21:20 crc kubenswrapper[4777]: I1124 17:21:20.708366 4777 scope.go:117] "RemoveContainer" containerID="8fff73dbee5b4eb854cbde38f25bb95e59a9bec9ffcf68c8fe6681e1bc1e3467" Nov 24 17:21:20 crc kubenswrapper[4777]: I1124 17:21:20.759084 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-k48pc" event={"ID":"6d2fef26-a53d-41dd-8523-b451f2ea8127","Type":"ContainerStarted","Data":"772aac6beaca59b436aba0c8bc944aad346788833a5922ff2ee7deaecb1cea4c"} Nov 24 17:21:20 crc kubenswrapper[4777]: I1124 17:21:20.762628 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5b3890d7-b149-4be6-bff8-60f8b7de9e22","Type":"ContainerStarted","Data":"72433a3f6fdb73cecb9f1ff2e4d32735952d2974d6c98d8033a7821ff9d83cc1"} Nov 24 17:21:20 crc kubenswrapper[4777]: E1124 17:21:20.803301 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current\\\"\"" pod="openstack/cloudkitty-db-sync-57hm5" podUID="a5ad0d8c-d098-4a37-a8f3-2563fff8b620" Nov 24 17:21:20 crc kubenswrapper[4777]: I1124 17:21:20.839459 4777 scope.go:117] "RemoveContainer" containerID="1740992d52d331d743216a96923b96754484801652f49b7638814c53dc620811" Nov 24 17:21:20 crc kubenswrapper[4777]: I1124 17:21:20.842474 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:21:20 crc kubenswrapper[4777]: I1124 17:21:20.954054 4777 scope.go:117] "RemoveContainer" containerID="5a67315ae43c2a6697812e5a7b6f9cc93c9ed13f3a104aeb98979a1cf9072dde" Nov 24 17:21:21 crc kubenswrapper[4777]: I1124 17:21:21.776925 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0eeb390-6ffc-41c8-a816-0b4454730e07","Type":"ContainerStarted","Data":"28cd9ebd6d1385ee8cc3d47a866c0c2c36fbee6fda63ce837a511c472a6be062"} Nov 24 17:21:21 crc kubenswrapper[4777]: I1124 17:21:21.779256 4777 generic.go:334] "Generic (PLEG): container finished" podID="b01f3e72-974a-4649-84a3-d23609d774a9" containerID="935312b5eed46e207be26dfa6e5cc349dffef7c3e3093aa5bf5591c49fdaa41b" exitCode=0 Nov 24 17:21:21 crc kubenswrapper[4777]: I1124 17:21:21.779301 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kl4mk" event={"ID":"b01f3e72-974a-4649-84a3-d23609d774a9","Type":"ContainerDied","Data":"935312b5eed46e207be26dfa6e5cc349dffef7c3e3093aa5bf5591c49fdaa41b"} Nov 24 17:21:21 crc kubenswrapper[4777]: I1124 17:21:21.782369 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-v7hr7" event={"ID":"d4151db5-e476-4ffa-a038-369874b439cc","Type":"ContainerStarted","Data":"ea7e7c17d4adfe7017b3c1ba21aa4f3026ea25523f283fc55a8ac4b0acecf7c6"} Nov 24 17:21:21 crc kubenswrapper[4777]: I1124 17:21:21.800650 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5b3890d7-b149-4be6-bff8-60f8b7de9e22","Type":"ContainerStarted","Data":"42977bcbcf9da3b11a6cd2bdd47f71900fec096419ad4bb2584edc54cb20e9be"} Nov 24 17:21:21 crc kubenswrapper[4777]: I1124 17:21:21.802424 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-k48pc" event={"ID":"6d2fef26-a53d-41dd-8523-b451f2ea8127","Type":"ContainerStarted","Data":"dc3a5cc17f9fd14cb69378682b5c09769c0067f44ee7ca44e40a7b0caa65a536"} Nov 24 17:21:21 crc kubenswrapper[4777]: I1124 17:21:21.812859 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pwl4w" event={"ID":"d84e561b-fba7-4438-9959-18e1b4bd889c","Type":"ContainerStarted","Data":"976ea245350feaf855dc1afa93656b1d0e6f42e14b8b5449334c3e0ea9661d25"} Nov 24 17:21:21 crc kubenswrapper[4777]: I1124 17:21:21.820954 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d03c2407-e0ad-49da-91a9-c9d3e8e67692","Type":"ContainerStarted","Data":"fdfe37b10df787951845c17e6d8025051d9dd87fe20c768b6d1a1a5b901f0777"} Nov 24 17:21:21 crc kubenswrapper[4777]: I1124 17:21:21.821025 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d03c2407-e0ad-49da-91a9-c9d3e8e67692","Type":"ContainerStarted","Data":"e2e6dceb91b23e0437e91788b4a923a59cadd2eccf302e3c4b729355fc4ee9b8"} Nov 24 17:21:21 crc kubenswrapper[4777]: I1124 17:21:21.821387 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-v7hr7" podStartSLOduration=3.514204607 podStartE2EDuration="42.821361291s" podCreationTimestamp="2025-11-24 17:20:39 +0000 UTC" firstStartedPulling="2025-11-24 17:20:41.385148263 +0000 UTC m=+1199.544343312" lastFinishedPulling="2025-11-24 17:21:20.692304947 +0000 UTC m=+1238.851499996" observedRunningTime="2025-11-24 17:21:21.812419743 +0000 UTC m=+1239.971614792" watchObservedRunningTime="2025-11-24 17:21:21.821361291 +0000 UTC m=+1239.980556340" Nov 24 17:21:21 crc kubenswrapper[4777]: I1124 17:21:21.834769 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-k48pc" podStartSLOduration=23.834749986 podStartE2EDuration="23.834749986s" podCreationTimestamp="2025-11-24 17:20:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:21:21.827922609 +0000 UTC m=+1239.987117668" watchObservedRunningTime="2025-11-24 17:21:21.834749986 +0000 UTC m=+1239.993945035" Nov 24 17:21:21 crc kubenswrapper[4777]: I1124 17:21:21.860112 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-pwl4w" podStartSLOduration=7.273346343 podStartE2EDuration="42.860091794s" podCreationTimestamp="2025-11-24 17:20:39 +0000 UTC" firstStartedPulling="2025-11-24 17:20:41.701776957 +0000 UTC m=+1199.860972006" lastFinishedPulling="2025-11-24 17:21:17.288522398 +0000 UTC m=+1235.447717457" observedRunningTime="2025-11-24 17:21:21.855398699 +0000 UTC m=+1240.014593748" watchObservedRunningTime="2025-11-24 17:21:21.860091794 +0000 UTC m=+1240.019286853" Nov 24 17:21:22 crc kubenswrapper[4777]: I1124 17:21:22.829981 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d03c2407-e0ad-49da-91a9-c9d3e8e67692","Type":"ContainerStarted","Data":"26f51bd0e101b47349c5d126e029079ae8159da608e1677a07fad76ed55d5e5c"} Nov 24 17:21:22 crc kubenswrapper[4777]: I1124 17:21:22.833050 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5b3890d7-b149-4be6-bff8-60f8b7de9e22","Type":"ContainerStarted","Data":"a230424fa31831eca45d0d49503ee0856c035c927587fe4e5f9f93e71fc722f0"} Nov 24 17:21:22 crc kubenswrapper[4777]: I1124 17:21:22.862564 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.862546068 podStartE2EDuration="8.862546068s" podCreationTimestamp="2025-11-24 17:21:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:21:22.853169458 +0000 UTC m=+1241.012364507" watchObservedRunningTime="2025-11-24 17:21:22.862546068 +0000 UTC m=+1241.021741107" Nov 24 17:21:22 crc kubenswrapper[4777]: I1124 17:21:22.885173 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=24.885150388 podStartE2EDuration="24.885150388s" podCreationTimestamp="2025-11-24 17:20:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:21:22.876095768 +0000 UTC m=+1241.035290827" watchObservedRunningTime="2025-11-24 17:21:22.885150388 +0000 UTC m=+1241.044345437" Nov 24 17:21:23 crc kubenswrapper[4777]: I1124 17:21:23.239604 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kl4mk" Nov 24 17:21:23 crc kubenswrapper[4777]: I1124 17:21:23.323178 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsv5x\" (UniqueName: \"kubernetes.io/projected/b01f3e72-974a-4649-84a3-d23609d774a9-kube-api-access-dsv5x\") pod \"b01f3e72-974a-4649-84a3-d23609d774a9\" (UID: \"b01f3e72-974a-4649-84a3-d23609d774a9\") " Nov 24 17:21:23 crc kubenswrapper[4777]: I1124 17:21:23.323249 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b01f3e72-974a-4649-84a3-d23609d774a9-config\") pod \"b01f3e72-974a-4649-84a3-d23609d774a9\" (UID: \"b01f3e72-974a-4649-84a3-d23609d774a9\") " Nov 24 17:21:23 crc kubenswrapper[4777]: I1124 17:21:23.323389 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b01f3e72-974a-4649-84a3-d23609d774a9-combined-ca-bundle\") pod \"b01f3e72-974a-4649-84a3-d23609d774a9\" (UID: \"b01f3e72-974a-4649-84a3-d23609d774a9\") " Nov 24 17:21:23 crc kubenswrapper[4777]: I1124 17:21:23.332293 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b01f3e72-974a-4649-84a3-d23609d774a9-kube-api-access-dsv5x" (OuterVolumeSpecName: "kube-api-access-dsv5x") pod "b01f3e72-974a-4649-84a3-d23609d774a9" (UID: "b01f3e72-974a-4649-84a3-d23609d774a9"). InnerVolumeSpecName "kube-api-access-dsv5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:21:23 crc kubenswrapper[4777]: I1124 17:21:23.357223 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b01f3e72-974a-4649-84a3-d23609d774a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b01f3e72-974a-4649-84a3-d23609d774a9" (UID: "b01f3e72-974a-4649-84a3-d23609d774a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:23 crc kubenswrapper[4777]: I1124 17:21:23.373146 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b01f3e72-974a-4649-84a3-d23609d774a9-config" (OuterVolumeSpecName: "config") pod "b01f3e72-974a-4649-84a3-d23609d774a9" (UID: "b01f3e72-974a-4649-84a3-d23609d774a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:23 crc kubenswrapper[4777]: I1124 17:21:23.425947 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsv5x\" (UniqueName: \"kubernetes.io/projected/b01f3e72-974a-4649-84a3-d23609d774a9-kube-api-access-dsv5x\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:23 crc kubenswrapper[4777]: I1124 17:21:23.426067 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b01f3e72-974a-4649-84a3-d23609d774a9-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:23 crc kubenswrapper[4777]: I1124 17:21:23.426256 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b01f3e72-974a-4649-84a3-d23609d774a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:23 crc kubenswrapper[4777]: I1124 17:21:23.843229 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kl4mk" Nov 24 17:21:23 crc kubenswrapper[4777]: I1124 17:21:23.845156 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kl4mk" event={"ID":"b01f3e72-974a-4649-84a3-d23609d774a9","Type":"ContainerDied","Data":"bfec8407a77a8a3403892f589135343ad44c3830e737fe4e4709879774ffc3fa"} Nov 24 17:21:23 crc kubenswrapper[4777]: I1124 17:21:23.845197 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfec8407a77a8a3403892f589135343ad44c3830e737fe4e4709879774ffc3fa" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.040162 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-8vztl"] Nov 24 17:21:24 crc kubenswrapper[4777]: E1124 17:21:24.040750 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b01f3e72-974a-4649-84a3-d23609d774a9" containerName="neutron-db-sync" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.040776 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="b01f3e72-974a-4649-84a3-d23609d774a9" containerName="neutron-db-sync" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.041067 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="b01f3e72-974a-4649-84a3-d23609d774a9" containerName="neutron-db-sync" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.042458 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.054825 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-8vztl"] Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.143941 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.144026 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-config\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.144099 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lq2r\" (UniqueName: \"kubernetes.io/projected/e8be48c9-46ce-4570-ab17-85835584b238-kube-api-access-4lq2r\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.145079 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-dns-svc\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.145112 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.145135 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.181412 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-77b7887964-hxn5r"] Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.183125 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.189700 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.189909 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.190052 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.190158 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-j8kfp" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.196836 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77b7887964-hxn5r"] Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.246987 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.247031 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-config\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.247075 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lq2r\" (UniqueName: \"kubernetes.io/projected/e8be48c9-46ce-4570-ab17-85835584b238-kube-api-access-4lq2r\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.247138 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-dns-svc\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.247157 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.247178 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.247940 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.248480 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.249160 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-config\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.249235 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-dns-svc\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.249605 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.277309 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lq2r\" (UniqueName: \"kubernetes.io/projected/e8be48c9-46ce-4570-ab17-85835584b238-kube-api-access-4lq2r\") pod \"dnsmasq-dns-6b7b667979-8vztl\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.348815 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmfjx\" (UniqueName: \"kubernetes.io/projected/6885283a-63d3-44a9-928c-39aff5a46e93-kube-api-access-wmfjx\") pod \"neutron-77b7887964-hxn5r\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.348904 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-httpd-config\") pod \"neutron-77b7887964-hxn5r\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.348947 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-combined-ca-bundle\") pod \"neutron-77b7887964-hxn5r\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.349090 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-ovndb-tls-certs\") pod \"neutron-77b7887964-hxn5r\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.349113 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-config\") pod \"neutron-77b7887964-hxn5r\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.373419 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.409207 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.409483 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.451054 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-ovndb-tls-certs\") pod \"neutron-77b7887964-hxn5r\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.451100 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-config\") pod \"neutron-77b7887964-hxn5r\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.451177 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmfjx\" (UniqueName: \"kubernetes.io/projected/6885283a-63d3-44a9-928c-39aff5a46e93-kube-api-access-wmfjx\") pod \"neutron-77b7887964-hxn5r\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.451207 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-httpd-config\") pod \"neutron-77b7887964-hxn5r\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.451234 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-combined-ca-bundle\") pod \"neutron-77b7887964-hxn5r\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.464209 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.464996 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-combined-ca-bundle\") pod \"neutron-77b7887964-hxn5r\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.465159 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-httpd-config\") pod \"neutron-77b7887964-hxn5r\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.466292 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-config\") pod \"neutron-77b7887964-hxn5r\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.467834 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmfjx\" (UniqueName: \"kubernetes.io/projected/6885283a-63d3-44a9-928c-39aff5a46e93-kube-api-access-wmfjx\") pod \"neutron-77b7887964-hxn5r\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.488670 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-ovndb-tls-certs\") pod \"neutron-77b7887964-hxn5r\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.513901 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.548609 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.853470 4777 generic.go:334] "Generic (PLEG): container finished" podID="6d2fef26-a53d-41dd-8523-b451f2ea8127" containerID="dc3a5cc17f9fd14cb69378682b5c09769c0067f44ee7ca44e40a7b0caa65a536" exitCode=0 Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.854542 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-k48pc" event={"ID":"6d2fef26-a53d-41dd-8523-b451f2ea8127","Type":"ContainerDied","Data":"dc3a5cc17f9fd14cb69378682b5c09769c0067f44ee7ca44e40a7b0caa65a536"} Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.857979 4777 generic.go:334] "Generic (PLEG): container finished" podID="d84e561b-fba7-4438-9959-18e1b4bd889c" containerID="976ea245350feaf855dc1afa93656b1d0e6f42e14b8b5449334c3e0ea9661d25" exitCode=0 Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.858917 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pwl4w" event={"ID":"d84e561b-fba7-4438-9959-18e1b4bd889c","Type":"ContainerDied","Data":"976ea245350feaf855dc1afa93656b1d0e6f42e14b8b5449334c3e0ea9661d25"} Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.859041 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 17:21:24 crc kubenswrapper[4777]: I1124 17:21:24.859182 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 17:21:25 crc kubenswrapper[4777]: I1124 17:21:25.868831 4777 generic.go:334] "Generic (PLEG): container finished" podID="d4151db5-e476-4ffa-a038-369874b439cc" containerID="ea7e7c17d4adfe7017b3c1ba21aa4f3026ea25523f283fc55a8ac4b0acecf7c6" exitCode=0 Nov 24 17:21:25 crc kubenswrapper[4777]: I1124 17:21:25.869060 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-v7hr7" event={"ID":"d4151db5-e476-4ffa-a038-369874b439cc","Type":"ContainerDied","Data":"ea7e7c17d4adfe7017b3c1ba21aa4f3026ea25523f283fc55a8ac4b0acecf7c6"} Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.340339 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5984748495-2rxn6"] Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.342080 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.344535 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.346692 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.352513 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5984748495-2rxn6"] Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.420620 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-ovndb-tls-certs\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.420667 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-public-tls-certs\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.420727 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-config\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.420799 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-httpd-config\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.420817 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-internal-tls-certs\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.420846 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-combined-ca-bundle\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.420868 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5nnp\" (UniqueName: \"kubernetes.io/projected/42e06676-5c5b-4684-b525-669d27d31818-kube-api-access-z5nnp\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.515465 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.520889 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pwl4w" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.523029 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-combined-ca-bundle\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.523079 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5nnp\" (UniqueName: \"kubernetes.io/projected/42e06676-5c5b-4684-b525-669d27d31818-kube-api-access-z5nnp\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.523204 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-ovndb-tls-certs\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.523237 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-public-tls-certs\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.523293 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-config\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.523352 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-httpd-config\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.523378 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-internal-tls-certs\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.528598 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-httpd-config\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.530354 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-ovndb-tls-certs\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.530751 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-public-tls-certs\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.530915 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-internal-tls-certs\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.535265 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-combined-ca-bundle\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.551066 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/42e06676-5c5b-4684-b525-669d27d31818-config\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.554364 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5nnp\" (UniqueName: \"kubernetes.io/projected/42e06676-5c5b-4684-b525-669d27d31818-kube-api-access-z5nnp\") pod \"neutron-5984748495-2rxn6\" (UID: \"42e06676-5c5b-4684-b525-669d27d31818\") " pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.624676 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-config-data\") pod \"6d2fef26-a53d-41dd-8523-b451f2ea8127\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.624761 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-combined-ca-bundle\") pod \"6d2fef26-a53d-41dd-8523-b451f2ea8127\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.624811 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-credential-keys\") pod \"6d2fef26-a53d-41dd-8523-b451f2ea8127\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.624876 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-scripts\") pod \"d84e561b-fba7-4438-9959-18e1b4bd889c\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.624895 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-combined-ca-bundle\") pod \"d84e561b-fba7-4438-9959-18e1b4bd889c\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.624930 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d84e561b-fba7-4438-9959-18e1b4bd889c-logs\") pod \"d84e561b-fba7-4438-9959-18e1b4bd889c\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.624987 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h27lh\" (UniqueName: \"kubernetes.io/projected/d84e561b-fba7-4438-9959-18e1b4bd889c-kube-api-access-h27lh\") pod \"d84e561b-fba7-4438-9959-18e1b4bd889c\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.625015 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-fernet-keys\") pod \"6d2fef26-a53d-41dd-8523-b451f2ea8127\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.625046 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-scripts\") pod \"6d2fef26-a53d-41dd-8523-b451f2ea8127\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.625062 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-config-data\") pod \"d84e561b-fba7-4438-9959-18e1b4bd889c\" (UID: \"d84e561b-fba7-4438-9959-18e1b4bd889c\") " Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.625108 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp6rp\" (UniqueName: \"kubernetes.io/projected/6d2fef26-a53d-41dd-8523-b451f2ea8127-kube-api-access-dp6rp\") pod \"6d2fef26-a53d-41dd-8523-b451f2ea8127\" (UID: \"6d2fef26-a53d-41dd-8523-b451f2ea8127\") " Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.625676 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d84e561b-fba7-4438-9959-18e1b4bd889c-logs" (OuterVolumeSpecName: "logs") pod "d84e561b-fba7-4438-9959-18e1b4bd889c" (UID: "d84e561b-fba7-4438-9959-18e1b4bd889c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.629242 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-scripts" (OuterVolumeSpecName: "scripts") pod "6d2fef26-a53d-41dd-8523-b451f2ea8127" (UID: "6d2fef26-a53d-41dd-8523-b451f2ea8127"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.629611 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6d2fef26-a53d-41dd-8523-b451f2ea8127" (UID: "6d2fef26-a53d-41dd-8523-b451f2ea8127"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.630860 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d84e561b-fba7-4438-9959-18e1b4bd889c-kube-api-access-h27lh" (OuterVolumeSpecName: "kube-api-access-h27lh") pod "d84e561b-fba7-4438-9959-18e1b4bd889c" (UID: "d84e561b-fba7-4438-9959-18e1b4bd889c"). InnerVolumeSpecName "kube-api-access-h27lh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.631021 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d2fef26-a53d-41dd-8523-b451f2ea8127-kube-api-access-dp6rp" (OuterVolumeSpecName: "kube-api-access-dp6rp") pod "6d2fef26-a53d-41dd-8523-b451f2ea8127" (UID: "6d2fef26-a53d-41dd-8523-b451f2ea8127"). InnerVolumeSpecName "kube-api-access-dp6rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.632955 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-scripts" (OuterVolumeSpecName: "scripts") pod "d84e561b-fba7-4438-9959-18e1b4bd889c" (UID: "d84e561b-fba7-4438-9959-18e1b4bd889c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.633725 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6d2fef26-a53d-41dd-8523-b451f2ea8127" (UID: "6d2fef26-a53d-41dd-8523-b451f2ea8127"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.649897 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d84e561b-fba7-4438-9959-18e1b4bd889c" (UID: "d84e561b-fba7-4438-9959-18e1b4bd889c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.652195 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-config-data" (OuterVolumeSpecName: "config-data") pod "6d2fef26-a53d-41dd-8523-b451f2ea8127" (UID: "6d2fef26-a53d-41dd-8523-b451f2ea8127"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.667712 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d2fef26-a53d-41dd-8523-b451f2ea8127" (UID: "6d2fef26-a53d-41dd-8523-b451f2ea8127"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.670468 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.674872 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-config-data" (OuterVolumeSpecName: "config-data") pod "d84e561b-fba7-4438-9959-18e1b4bd889c" (UID: "d84e561b-fba7-4438-9959-18e1b4bd889c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.726813 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h27lh\" (UniqueName: \"kubernetes.io/projected/d84e561b-fba7-4438-9959-18e1b4bd889c-kube-api-access-h27lh\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.726850 4777 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.726861 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.726869 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.726877 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp6rp\" (UniqueName: \"kubernetes.io/projected/6d2fef26-a53d-41dd-8523-b451f2ea8127-kube-api-access-dp6rp\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.726887 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.726895 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.726903 4777 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6d2fef26-a53d-41dd-8523-b451f2ea8127-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.726910 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.726918 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84e561b-fba7-4438-9959-18e1b4bd889c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.726925 4777 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d84e561b-fba7-4438-9959-18e1b4bd889c-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.884221 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pwl4w" event={"ID":"d84e561b-fba7-4438-9959-18e1b4bd889c","Type":"ContainerDied","Data":"6aa2f3918b726f9c1286241aa1f0cce42f1e0d7e5a6945a0ae798b2b76b59ecc"} Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.884267 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6aa2f3918b726f9c1286241aa1f0cce42f1e0d7e5a6945a0ae798b2b76b59ecc" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.884354 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pwl4w" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.894689 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-k48pc" Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.895327 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-k48pc" event={"ID":"6d2fef26-a53d-41dd-8523-b451f2ea8127","Type":"ContainerDied","Data":"772aac6beaca59b436aba0c8bc944aad346788833a5922ff2ee7deaecb1cea4c"} Nov 24 17:21:26 crc kubenswrapper[4777]: I1124 17:21:26.895368 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="772aac6beaca59b436aba0c8bc944aad346788833a5922ff2ee7deaecb1cea4c" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.050053 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5877679d68-q5nlj"] Nov 24 17:21:27 crc kubenswrapper[4777]: E1124 17:21:27.050490 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d2fef26-a53d-41dd-8523-b451f2ea8127" containerName="keystone-bootstrap" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.050505 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d2fef26-a53d-41dd-8523-b451f2ea8127" containerName="keystone-bootstrap" Nov 24 17:21:27 crc kubenswrapper[4777]: E1124 17:21:27.050536 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d84e561b-fba7-4438-9959-18e1b4bd889c" containerName="placement-db-sync" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.050543 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="d84e561b-fba7-4438-9959-18e1b4bd889c" containerName="placement-db-sync" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.050704 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d2fef26-a53d-41dd-8523-b451f2ea8127" containerName="keystone-bootstrap" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.050733 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="d84e561b-fba7-4438-9959-18e1b4bd889c" containerName="placement-db-sync" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.051418 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.054274 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.054496 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vzrxj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.054621 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.054736 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.054885 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.055107 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.081228 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5877679d68-q5nlj"] Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.098154 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-55f98d7b-5njfr"] Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.100312 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.106115 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.106164 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.106384 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-44v99" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.106619 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.107421 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.155132 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-public-tls-certs\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.155236 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-config-data\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.155344 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-scripts\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.155395 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlrbj\" (UniqueName: \"kubernetes.io/projected/8021c40d-848c-4365-84f2-9878b03d4656-kube-api-access-zlrbj\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.155430 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-fernet-keys\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.155562 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-credential-keys\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.155587 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-internal-tls-certs\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.155639 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-combined-ca-bundle\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.155821 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-55f98d7b-5njfr"] Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.259541 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-public-tls-certs\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.259612 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-logs\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.259774 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-credential-keys\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.259819 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-internal-tls-certs\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.259875 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-internal-tls-certs\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.259953 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-combined-ca-bundle\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.260254 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-scripts\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.260284 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-public-tls-certs\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.260353 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-combined-ca-bundle\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.260768 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-config-data\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.261345 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zvdq\" (UniqueName: \"kubernetes.io/projected/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-kube-api-access-2zvdq\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.261367 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-config-data\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.261507 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-scripts\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.261640 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlrbj\" (UniqueName: \"kubernetes.io/projected/8021c40d-848c-4365-84f2-9878b03d4656-kube-api-access-zlrbj\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.261735 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-fernet-keys\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.265698 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-combined-ca-bundle\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.267170 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-internal-tls-certs\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.267536 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-config-data\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.268402 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-credential-keys\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.270465 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-public-tls-certs\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.275234 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-fernet-keys\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.275508 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8021c40d-848c-4365-84f2-9878b03d4656-scripts\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.279731 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlrbj\" (UniqueName: \"kubernetes.io/projected/8021c40d-848c-4365-84f2-9878b03d4656-kube-api-access-zlrbj\") pod \"keystone-5877679d68-q5nlj\" (UID: \"8021c40d-848c-4365-84f2-9878b03d4656\") " pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.364590 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-scripts\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.364661 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-combined-ca-bundle\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.364718 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zvdq\" (UniqueName: \"kubernetes.io/projected/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-kube-api-access-2zvdq\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.364741 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-config-data\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.364853 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-public-tls-certs\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.364892 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-logs\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.364943 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-internal-tls-certs\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.366842 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-logs\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.369469 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-scripts\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.370291 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-internal-tls-certs\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.370954 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-combined-ca-bundle\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.375191 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-public-tls-certs\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.376155 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-config-data\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.381822 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zvdq\" (UniqueName: \"kubernetes.io/projected/b25fa9b0-aec5-4a33-aefe-2ee5685d0e88-kube-api-access-2zvdq\") pod \"placement-55f98d7b-5njfr\" (UID: \"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88\") " pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.409899 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:27 crc kubenswrapper[4777]: I1124 17:21:27.444480 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:28 crc kubenswrapper[4777]: I1124 17:21:28.851641 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 17:21:28 crc kubenswrapper[4777]: I1124 17:21:28.952128 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 17:21:28 crc kubenswrapper[4777]: I1124 17:21:28.952172 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 17:21:28 crc kubenswrapper[4777]: I1124 17:21:28.952183 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 17:21:28 crc kubenswrapper[4777]: I1124 17:21:28.952192 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 17:21:29 crc kubenswrapper[4777]: I1124 17:21:29.003668 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 17:21:29 crc kubenswrapper[4777]: I1124 17:21:29.005027 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 17:21:30 crc kubenswrapper[4777]: I1124 17:21:30.825549 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 17:21:32 crc kubenswrapper[4777]: I1124 17:21:32.177755 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 17:21:32 crc kubenswrapper[4777]: I1124 17:21:32.178240 4777 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 17:21:32 crc kubenswrapper[4777]: I1124 17:21:32.181613 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 17:21:37 crc kubenswrapper[4777]: I1124 17:21:37.429918 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-v7hr7" Nov 24 17:21:37 crc kubenswrapper[4777]: I1124 17:21:37.593583 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d4151db5-e476-4ffa-a038-369874b439cc-db-sync-config-data\") pod \"d4151db5-e476-4ffa-a038-369874b439cc\" (UID: \"d4151db5-e476-4ffa-a038-369874b439cc\") " Nov 24 17:21:37 crc kubenswrapper[4777]: I1124 17:21:37.593972 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rszcl\" (UniqueName: \"kubernetes.io/projected/d4151db5-e476-4ffa-a038-369874b439cc-kube-api-access-rszcl\") pod \"d4151db5-e476-4ffa-a038-369874b439cc\" (UID: \"d4151db5-e476-4ffa-a038-369874b439cc\") " Nov 24 17:21:37 crc kubenswrapper[4777]: I1124 17:21:37.594088 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4151db5-e476-4ffa-a038-369874b439cc-combined-ca-bundle\") pod \"d4151db5-e476-4ffa-a038-369874b439cc\" (UID: \"d4151db5-e476-4ffa-a038-369874b439cc\") " Nov 24 17:21:37 crc kubenswrapper[4777]: I1124 17:21:37.600086 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4151db5-e476-4ffa-a038-369874b439cc-kube-api-access-rszcl" (OuterVolumeSpecName: "kube-api-access-rszcl") pod "d4151db5-e476-4ffa-a038-369874b439cc" (UID: "d4151db5-e476-4ffa-a038-369874b439cc"). InnerVolumeSpecName "kube-api-access-rszcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:21:37 crc kubenswrapper[4777]: I1124 17:21:37.600389 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4151db5-e476-4ffa-a038-369874b439cc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d4151db5-e476-4ffa-a038-369874b439cc" (UID: "d4151db5-e476-4ffa-a038-369874b439cc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:37 crc kubenswrapper[4777]: I1124 17:21:37.625168 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4151db5-e476-4ffa-a038-369874b439cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4151db5-e476-4ffa-a038-369874b439cc" (UID: "d4151db5-e476-4ffa-a038-369874b439cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:37 crc kubenswrapper[4777]: I1124 17:21:37.696252 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rszcl\" (UniqueName: \"kubernetes.io/projected/d4151db5-e476-4ffa-a038-369874b439cc-kube-api-access-rszcl\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:37 crc kubenswrapper[4777]: I1124 17:21:37.696295 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4151db5-e476-4ffa-a038-369874b439cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:37 crc kubenswrapper[4777]: I1124 17:21:37.696306 4777 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d4151db5-e476-4ffa-a038-369874b439cc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:37 crc kubenswrapper[4777]: I1124 17:21:37.833281 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-8vztl"] Nov 24 17:21:37 crc kubenswrapper[4777]: W1124 17:21:37.949442 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb25fa9b0_aec5_4a33_aefe_2ee5685d0e88.slice/crio-2755318acc53c8341f024b749ccd9025c27e84e23a2976bf8a2ffe2037311371 WatchSource:0}: Error finding container 2755318acc53c8341f024b749ccd9025c27e84e23a2976bf8a2ffe2037311371: Status 404 returned error can't find the container with id 2755318acc53c8341f024b749ccd9025c27e84e23a2976bf8a2ffe2037311371 Nov 24 17:21:37 crc kubenswrapper[4777]: I1124 17:21:37.957162 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-55f98d7b-5njfr"] Nov 24 17:21:37 crc kubenswrapper[4777]: I1124 17:21:37.969508 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5877679d68-q5nlj"] Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.017319 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-8vztl" event={"ID":"e8be48c9-46ce-4570-ab17-85835584b238","Type":"ContainerStarted","Data":"1d29f579fd566a50927673a4474abe20a258ca0410cd1d510f32f6e83a6b8260"} Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.024717 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-v7hr7" event={"ID":"d4151db5-e476-4ffa-a038-369874b439cc","Type":"ContainerDied","Data":"f309353a1a7183b0c92b7a6290f177bbd2df9f7096e08d70d4c6df4f7392333e"} Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.024780 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f309353a1a7183b0c92b7a6290f177bbd2df9f7096e08d70d4c6df4f7392333e" Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.024862 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-v7hr7" Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.029639 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-55f98d7b-5njfr" event={"ID":"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88","Type":"ContainerStarted","Data":"2755318acc53c8341f024b749ccd9025c27e84e23a2976bf8a2ffe2037311371"} Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.034164 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5877679d68-q5nlj" event={"ID":"8021c40d-848c-4365-84f2-9878b03d4656","Type":"ContainerStarted","Data":"bb49b5dbc65bbe71ac49b2db6693208ffe7f84d15fde36fce7b96ea22b88d9aa"} Nov 24 17:21:38 crc kubenswrapper[4777]: W1124 17:21:38.048858 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42e06676_5c5b_4684_b525_669d27d31818.slice/crio-458e2aa758d3997335d92e775462620c05b374d02b100d2f0d3ff18f03a64026 WatchSource:0}: Error finding container 458e2aa758d3997335d92e775462620c05b374d02b100d2f0d3ff18f03a64026: Status 404 returned error can't find the container with id 458e2aa758d3997335d92e775462620c05b374d02b100d2f0d3ff18f03a64026 Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.058543 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5984748495-2rxn6"] Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.892086 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5d57df9989-fbq7p"] Nov 24 17:21:38 crc kubenswrapper[4777]: E1124 17:21:38.893032 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4151db5-e476-4ffa-a038-369874b439cc" containerName="barbican-db-sync" Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.893047 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4151db5-e476-4ffa-a038-369874b439cc" containerName="barbican-db-sync" Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.893256 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4151db5-e476-4ffa-a038-369874b439cc" containerName="barbican-db-sync" Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.894289 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.897125 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.898573 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.899006 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-f5rnw" Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.909087 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-845875d46-n8hgg"] Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.910930 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.912911 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.927575 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5d57df9989-fbq7p"] Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.952484 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77b7887964-hxn5r"] Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.966311 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-8vztl"] Nov 24 17:21:38 crc kubenswrapper[4777]: I1124 17:21:38.987226 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-845875d46-n8hgg"] Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.002244 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-hngmc"] Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.004106 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.021036 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-hngmc"] Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.023874 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dded6598-74bf-4395-88e0-37152b4dd16c-logs\") pod \"barbican-worker-5d57df9989-fbq7p\" (UID: \"dded6598-74bf-4395-88e0-37152b4dd16c\") " pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.023916 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx8bq\" (UniqueName: \"kubernetes.io/projected/dded6598-74bf-4395-88e0-37152b4dd16c-kube-api-access-mx8bq\") pod \"barbican-worker-5d57df9989-fbq7p\" (UID: \"dded6598-74bf-4395-88e0-37152b4dd16c\") " pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.027702 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d173b040-b120-4c77-ba3c-b13c41a6ac37-config-data-custom\") pod \"barbican-keystone-listener-845875d46-n8hgg\" (UID: \"d173b040-b120-4c77-ba3c-b13c41a6ac37\") " pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.027735 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dded6598-74bf-4395-88e0-37152b4dd16c-config-data-custom\") pod \"barbican-worker-5d57df9989-fbq7p\" (UID: \"dded6598-74bf-4395-88e0-37152b4dd16c\") " pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.027795 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khmlr\" (UniqueName: \"kubernetes.io/projected/d173b040-b120-4c77-ba3c-b13c41a6ac37-kube-api-access-khmlr\") pod \"barbican-keystone-listener-845875d46-n8hgg\" (UID: \"d173b040-b120-4c77-ba3c-b13c41a6ac37\") " pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.027827 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dded6598-74bf-4395-88e0-37152b4dd16c-combined-ca-bundle\") pod \"barbican-worker-5d57df9989-fbq7p\" (UID: \"dded6598-74bf-4395-88e0-37152b4dd16c\") " pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.027872 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d173b040-b120-4c77-ba3c-b13c41a6ac37-logs\") pod \"barbican-keystone-listener-845875d46-n8hgg\" (UID: \"d173b040-b120-4c77-ba3c-b13c41a6ac37\") " pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.027897 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dded6598-74bf-4395-88e0-37152b4dd16c-config-data\") pod \"barbican-worker-5d57df9989-fbq7p\" (UID: \"dded6598-74bf-4395-88e0-37152b4dd16c\") " pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.027957 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d173b040-b120-4c77-ba3c-b13c41a6ac37-config-data\") pod \"barbican-keystone-listener-845875d46-n8hgg\" (UID: \"d173b040-b120-4c77-ba3c-b13c41a6ac37\") " pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.027990 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d173b040-b120-4c77-ba3c-b13c41a6ac37-combined-ca-bundle\") pod \"barbican-keystone-listener-845875d46-n8hgg\" (UID: \"d173b040-b120-4c77-ba3c-b13c41a6ac37\") " pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.083803 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77b7887964-hxn5r" event={"ID":"6885283a-63d3-44a9-928c-39aff5a46e93","Type":"ContainerStarted","Data":"33b45c448ba260621ea3e3604ef025e97f509e4fce71058422ef02be4aa1ad03"} Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.086195 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5984748495-2rxn6" event={"ID":"42e06676-5c5b-4684-b525-669d27d31818","Type":"ContainerStarted","Data":"458e2aa758d3997335d92e775462620c05b374d02b100d2f0d3ff18f03a64026"} Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.130777 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.130837 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d173b040-b120-4c77-ba3c-b13c41a6ac37-logs\") pod \"barbican-keystone-listener-845875d46-n8hgg\" (UID: \"d173b040-b120-4c77-ba3c-b13c41a6ac37\") " pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.130861 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dded6598-74bf-4395-88e0-37152b4dd16c-config-data\") pod \"barbican-worker-5d57df9989-fbq7p\" (UID: \"dded6598-74bf-4395-88e0-37152b4dd16c\") " pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.130883 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-config\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.130918 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d173b040-b120-4c77-ba3c-b13c41a6ac37-config-data\") pod \"barbican-keystone-listener-845875d46-n8hgg\" (UID: \"d173b040-b120-4c77-ba3c-b13c41a6ac37\") " pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.130938 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d173b040-b120-4c77-ba3c-b13c41a6ac37-combined-ca-bundle\") pod \"barbican-keystone-listener-845875d46-n8hgg\" (UID: \"d173b040-b120-4c77-ba3c-b13c41a6ac37\") " pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.130985 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z54wf\" (UniqueName: \"kubernetes.io/projected/b1602c3f-a6a0-40b7-a56b-df49845a3242-kube-api-access-z54wf\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.131004 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dded6598-74bf-4395-88e0-37152b4dd16c-logs\") pod \"barbican-worker-5d57df9989-fbq7p\" (UID: \"dded6598-74bf-4395-88e0-37152b4dd16c\") " pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.131023 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx8bq\" (UniqueName: \"kubernetes.io/projected/dded6598-74bf-4395-88e0-37152b4dd16c-kube-api-access-mx8bq\") pod \"barbican-worker-5d57df9989-fbq7p\" (UID: \"dded6598-74bf-4395-88e0-37152b4dd16c\") " pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.131058 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.131083 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.131102 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d173b040-b120-4c77-ba3c-b13c41a6ac37-config-data-custom\") pod \"barbican-keystone-listener-845875d46-n8hgg\" (UID: \"d173b040-b120-4c77-ba3c-b13c41a6ac37\") " pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.131120 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dded6598-74bf-4395-88e0-37152b4dd16c-config-data-custom\") pod \"barbican-worker-5d57df9989-fbq7p\" (UID: \"dded6598-74bf-4395-88e0-37152b4dd16c\") " pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.131150 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.131180 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khmlr\" (UniqueName: \"kubernetes.io/projected/d173b040-b120-4c77-ba3c-b13c41a6ac37-kube-api-access-khmlr\") pod \"barbican-keystone-listener-845875d46-n8hgg\" (UID: \"d173b040-b120-4c77-ba3c-b13c41a6ac37\") " pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.131204 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dded6598-74bf-4395-88e0-37152b4dd16c-combined-ca-bundle\") pod \"barbican-worker-5d57df9989-fbq7p\" (UID: \"dded6598-74bf-4395-88e0-37152b4dd16c\") " pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.131591 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dded6598-74bf-4395-88e0-37152b4dd16c-logs\") pod \"barbican-worker-5d57df9989-fbq7p\" (UID: \"dded6598-74bf-4395-88e0-37152b4dd16c\") " pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.131938 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d173b040-b120-4c77-ba3c-b13c41a6ac37-logs\") pod \"barbican-keystone-listener-845875d46-n8hgg\" (UID: \"d173b040-b120-4c77-ba3c-b13c41a6ac37\") " pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.135217 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dded6598-74bf-4395-88e0-37152b4dd16c-combined-ca-bundle\") pod \"barbican-worker-5d57df9989-fbq7p\" (UID: \"dded6598-74bf-4395-88e0-37152b4dd16c\") " pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.137878 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d173b040-b120-4c77-ba3c-b13c41a6ac37-config-data-custom\") pod \"barbican-keystone-listener-845875d46-n8hgg\" (UID: \"d173b040-b120-4c77-ba3c-b13c41a6ac37\") " pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.145537 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d173b040-b120-4c77-ba3c-b13c41a6ac37-config-data\") pod \"barbican-keystone-listener-845875d46-n8hgg\" (UID: \"d173b040-b120-4c77-ba3c-b13c41a6ac37\") " pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.145857 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d173b040-b120-4c77-ba3c-b13c41a6ac37-combined-ca-bundle\") pod \"barbican-keystone-listener-845875d46-n8hgg\" (UID: \"d173b040-b120-4c77-ba3c-b13c41a6ac37\") " pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.146477 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dded6598-74bf-4395-88e0-37152b4dd16c-config-data-custom\") pod \"barbican-worker-5d57df9989-fbq7p\" (UID: \"dded6598-74bf-4395-88e0-37152b4dd16c\") " pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.147207 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dded6598-74bf-4395-88e0-37152b4dd16c-config-data\") pod \"barbican-worker-5d57df9989-fbq7p\" (UID: \"dded6598-74bf-4395-88e0-37152b4dd16c\") " pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.160358 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx8bq\" (UniqueName: \"kubernetes.io/projected/dded6598-74bf-4395-88e0-37152b4dd16c-kube-api-access-mx8bq\") pod \"barbican-worker-5d57df9989-fbq7p\" (UID: \"dded6598-74bf-4395-88e0-37152b4dd16c\") " pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.169330 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khmlr\" (UniqueName: \"kubernetes.io/projected/d173b040-b120-4c77-ba3c-b13c41a6ac37-kube-api-access-khmlr\") pod \"barbican-keystone-listener-845875d46-n8hgg\" (UID: \"d173b040-b120-4c77-ba3c-b13c41a6ac37\") " pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.172753 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-79b9d8bcdd-vfnwm"] Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.174342 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.176918 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.182066 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-845875d46-n8hgg" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.191488 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-79b9d8bcdd-vfnwm"] Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.232827 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.233151 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.233181 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-config\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.233236 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z54wf\" (UniqueName: \"kubernetes.io/projected/b1602c3f-a6a0-40b7-a56b-df49845a3242-kube-api-access-z54wf\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.233276 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.233307 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.234244 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.234769 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.235608 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.238180 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-config\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.238357 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.277792 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z54wf\" (UniqueName: \"kubernetes.io/projected/b1602c3f-a6a0-40b7-a56b-df49845a3242-kube-api-access-z54wf\") pod \"dnsmasq-dns-848cf88cfc-hngmc\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.335928 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwvfm\" (UniqueName: \"kubernetes.io/projected/33638f0f-d115-472d-8de8-3d459bc6c19c-kube-api-access-pwvfm\") pod \"barbican-api-79b9d8bcdd-vfnwm\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.336001 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-config-data\") pod \"barbican-api-79b9d8bcdd-vfnwm\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.336044 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-config-data-custom\") pod \"barbican-api-79b9d8bcdd-vfnwm\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.336169 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33638f0f-d115-472d-8de8-3d459bc6c19c-logs\") pod \"barbican-api-79b9d8bcdd-vfnwm\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.336241 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-combined-ca-bundle\") pod \"barbican-api-79b9d8bcdd-vfnwm\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.430477 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5d57df9989-fbq7p" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.436132 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.440000 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-combined-ca-bundle\") pod \"barbican-api-79b9d8bcdd-vfnwm\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.440047 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwvfm\" (UniqueName: \"kubernetes.io/projected/33638f0f-d115-472d-8de8-3d459bc6c19c-kube-api-access-pwvfm\") pod \"barbican-api-79b9d8bcdd-vfnwm\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.440067 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-config-data\") pod \"barbican-api-79b9d8bcdd-vfnwm\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.440103 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-config-data-custom\") pod \"barbican-api-79b9d8bcdd-vfnwm\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.440223 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33638f0f-d115-472d-8de8-3d459bc6c19c-logs\") pod \"barbican-api-79b9d8bcdd-vfnwm\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.441122 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33638f0f-d115-472d-8de8-3d459bc6c19c-logs\") pod \"barbican-api-79b9d8bcdd-vfnwm\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.453591 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-config-data-custom\") pod \"barbican-api-79b9d8bcdd-vfnwm\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.470543 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-combined-ca-bundle\") pod \"barbican-api-79b9d8bcdd-vfnwm\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.485365 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-config-data\") pod \"barbican-api-79b9d8bcdd-vfnwm\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.490140 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwvfm\" (UniqueName: \"kubernetes.io/projected/33638f0f-d115-472d-8de8-3d459bc6c19c-kube-api-access-pwvfm\") pod \"barbican-api-79b9d8bcdd-vfnwm\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.582483 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:39 crc kubenswrapper[4777]: I1124 17:21:39.852393 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-845875d46-n8hgg"] Nov 24 17:21:40 crc kubenswrapper[4777]: I1124 17:21:40.083679 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5d57df9989-fbq7p"] Nov 24 17:21:40 crc kubenswrapper[4777]: I1124 17:21:40.106303 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5d57df9989-fbq7p" event={"ID":"dded6598-74bf-4395-88e0-37152b4dd16c","Type":"ContainerStarted","Data":"97c6c0e1792a8a692ce207d7374bd23a93e907d4ac3557e2f494d0801d2aea80"} Nov 24 17:21:40 crc kubenswrapper[4777]: I1124 17:21:40.110281 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-845875d46-n8hgg" event={"ID":"d173b040-b120-4c77-ba3c-b13c41a6ac37","Type":"ContainerStarted","Data":"b65367e56719b29ef97e18b80df8f4d7ae1f2156cb160fe16e567d9f04051285"} Nov 24 17:21:40 crc kubenswrapper[4777]: I1124 17:21:40.195904 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-hngmc"] Nov 24 17:21:40 crc kubenswrapper[4777]: W1124 17:21:40.199461 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1602c3f_a6a0_40b7_a56b_df49845a3242.slice/crio-4c7d48553db4b7f018904050faf1381e28c0bd4fe080df9b3f46e7d7b3f170b4 WatchSource:0}: Error finding container 4c7d48553db4b7f018904050faf1381e28c0bd4fe080df9b3f46e7d7b3f170b4: Status 404 returned error can't find the container with id 4c7d48553db4b7f018904050faf1381e28c0bd4fe080df9b3f46e7d7b3f170b4 Nov 24 17:21:40 crc kubenswrapper[4777]: I1124 17:21:40.211134 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-79b9d8bcdd-vfnwm"] Nov 24 17:21:40 crc kubenswrapper[4777]: E1124 17:21:40.389951 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/sg-core:latest" Nov 24 17:21:40 crc kubenswrapper[4777]: E1124 17:21:40.390191 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:sg-core,Image:quay.io/openstack-k8s-operators/sg-core:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:sg-core-conf-yaml,ReadOnly:false,MountPath:/etc/sg-core.conf.yaml,SubPath:sg-core.conf.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qd7sx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(c0eeb390-6ffc-41c8-a816-0b4454730e07): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.123819 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77b7887964-hxn5r" event={"ID":"6885283a-63d3-44a9-928c-39aff5a46e93","Type":"ContainerStarted","Data":"b49906183d185d99ed7bf30b881efa446b8a6139d8008b6ee5e4218e587e7324"} Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.124146 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77b7887964-hxn5r" event={"ID":"6885283a-63d3-44a9-928c-39aff5a46e93","Type":"ContainerStarted","Data":"efdb2752b15ceedc4b1933eabf81a88f9b4c0110a61f8e05eb3f4dbe01cd537c"} Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.124457 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.126216 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" event={"ID":"33638f0f-d115-472d-8de8-3d459bc6c19c","Type":"ContainerStarted","Data":"80a456e5941ed70510ff0ae6327a8bc1cea6d791e29114eeee24f07063cf8411"} Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.126257 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" event={"ID":"33638f0f-d115-472d-8de8-3d459bc6c19c","Type":"ContainerStarted","Data":"3b101f47418213fea25f858a695a939559d4f08001662bf6f2046cc56328b4e7"} Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.128625 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-55f98d7b-5njfr" event={"ID":"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88","Type":"ContainerStarted","Data":"f16e3eb5a8182deb258ae0d3f88018607ddc26a89836c669881ae21b95b6db5f"} Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.128661 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-55f98d7b-5njfr" event={"ID":"b25fa9b0-aec5-4a33-aefe-2ee5685d0e88","Type":"ContainerStarted","Data":"f976eaffd340e18d64b5c944b5d3afd1e92ce7857ac5fee743247c4524d38d7b"} Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.129581 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.129611 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.131463 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5877679d68-q5nlj" event={"ID":"8021c40d-848c-4365-84f2-9878b03d4656","Type":"ContainerStarted","Data":"af1663cadb26e6324656ec02a3f1e6c4592686066d9f00fc6744e114286e552e"} Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.131691 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.133073 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" event={"ID":"b1602c3f-a6a0-40b7-a56b-df49845a3242","Type":"ContainerStarted","Data":"dae56e2bd4581eae9f6c3e1176902d8795dcc89ece8c2d14f46d371fc437f64a"} Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.133104 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" event={"ID":"b1602c3f-a6a0-40b7-a56b-df49845a3242","Type":"ContainerStarted","Data":"4c7d48553db4b7f018904050faf1381e28c0bd4fe080df9b3f46e7d7b3f170b4"} Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.135557 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5984748495-2rxn6" event={"ID":"42e06676-5c5b-4684-b525-669d27d31818","Type":"ContainerStarted","Data":"68ee27a11006a7be6e18cc7a3ccb3b4c12b26b210b8c60208a2b7401b7bead3a"} Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.135600 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5984748495-2rxn6" event={"ID":"42e06676-5c5b-4684-b525-669d27d31818","Type":"ContainerStarted","Data":"049dbf50afadab9bf37355466aad0d30d7cf6f20303586adfa4f37732f811130"} Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.135659 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.138770 4777 generic.go:334] "Generic (PLEG): container finished" podID="e8be48c9-46ce-4570-ab17-85835584b238" containerID="229e8e158f9067960deac2fc2a75afcdfbddbe08930500f48b2e2d41562c0d9e" exitCode=0 Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.138807 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-8vztl" event={"ID":"e8be48c9-46ce-4570-ab17-85835584b238","Type":"ContainerDied","Data":"229e8e158f9067960deac2fc2a75afcdfbddbe08930500f48b2e2d41562c0d9e"} Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.149839 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-77b7887964-hxn5r" podStartSLOduration=17.14981112 podStartE2EDuration="17.14981112s" podCreationTimestamp="2025-11-24 17:21:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:21:41.143088576 +0000 UTC m=+1259.302283625" watchObservedRunningTime="2025-11-24 17:21:41.14981112 +0000 UTC m=+1259.309006169" Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.171940 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5877679d68-q5nlj" podStartSLOduration=15.171921855 podStartE2EDuration="15.171921855s" podCreationTimestamp="2025-11-24 17:21:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:21:41.161463915 +0000 UTC m=+1259.320658964" watchObservedRunningTime="2025-11-24 17:21:41.171921855 +0000 UTC m=+1259.331116894" Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.233717 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-55f98d7b-5njfr" podStartSLOduration=14.233703042 podStartE2EDuration="14.233703042s" podCreationTimestamp="2025-11-24 17:21:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:21:41.210369161 +0000 UTC m=+1259.369564220" watchObservedRunningTime="2025-11-24 17:21:41.233703042 +0000 UTC m=+1259.392898091" Nov 24 17:21:41 crc kubenswrapper[4777]: I1124 17:21:41.279747 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5984748495-2rxn6" podStartSLOduration=15.279732815 podStartE2EDuration="15.279732815s" podCreationTimestamp="2025-11-24 17:21:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:21:41.273445754 +0000 UTC m=+1259.432640803" watchObservedRunningTime="2025-11-24 17:21:41.279732815 +0000 UTC m=+1259.438927864" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.036753 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.142642 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-ovsdbserver-nb\") pod \"e8be48c9-46ce-4570-ab17-85835584b238\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.142694 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-config\") pod \"e8be48c9-46ce-4570-ab17-85835584b238\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.142751 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lq2r\" (UniqueName: \"kubernetes.io/projected/e8be48c9-46ce-4570-ab17-85835584b238-kube-api-access-4lq2r\") pod \"e8be48c9-46ce-4570-ab17-85835584b238\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.142842 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-dns-svc\") pod \"e8be48c9-46ce-4570-ab17-85835584b238\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.142909 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-ovsdbserver-sb\") pod \"e8be48c9-46ce-4570-ab17-85835584b238\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.142985 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-dns-swift-storage-0\") pod \"e8be48c9-46ce-4570-ab17-85835584b238\" (UID: \"e8be48c9-46ce-4570-ab17-85835584b238\") " Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.153901 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8be48c9-46ce-4570-ab17-85835584b238-kube-api-access-4lq2r" (OuterVolumeSpecName: "kube-api-access-4lq2r") pod "e8be48c9-46ce-4570-ab17-85835584b238" (UID: "e8be48c9-46ce-4570-ab17-85835584b238"). InnerVolumeSpecName "kube-api-access-4lq2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.163419 4777 generic.go:334] "Generic (PLEG): container finished" podID="b1602c3f-a6a0-40b7-a56b-df49845a3242" containerID="dae56e2bd4581eae9f6c3e1176902d8795dcc89ece8c2d14f46d371fc437f64a" exitCode=0 Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.163487 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" event={"ID":"b1602c3f-a6a0-40b7-a56b-df49845a3242","Type":"ContainerDied","Data":"dae56e2bd4581eae9f6c3e1176902d8795dcc89ece8c2d14f46d371fc437f64a"} Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.176313 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e8be48c9-46ce-4570-ab17-85835584b238" (UID: "e8be48c9-46ce-4570-ab17-85835584b238"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.182108 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-8vztl" event={"ID":"e8be48c9-46ce-4570-ab17-85835584b238","Type":"ContainerDied","Data":"1d29f579fd566a50927673a4474abe20a258ca0410cd1d510f32f6e83a6b8260"} Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.182160 4777 scope.go:117] "RemoveContainer" containerID="229e8e158f9067960deac2fc2a75afcdfbddbe08930500f48b2e2d41562c0d9e" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.183021 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-8vztl" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.188185 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e8be48c9-46ce-4570-ab17-85835584b238" (UID: "e8be48c9-46ce-4570-ab17-85835584b238"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.203942 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e8be48c9-46ce-4570-ab17-85835584b238" (UID: "e8be48c9-46ce-4570-ab17-85835584b238"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.207865 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e8be48c9-46ce-4570-ab17-85835584b238" (UID: "e8be48c9-46ce-4570-ab17-85835584b238"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.219487 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-config" (OuterVolumeSpecName: "config") pod "e8be48c9-46ce-4570-ab17-85835584b238" (UID: "e8be48c9-46ce-4570-ab17-85835584b238"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.245692 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.245720 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.245730 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lq2r\" (UniqueName: \"kubernetes.io/projected/e8be48c9-46ce-4570-ab17-85835584b238-kube-api-access-4lq2r\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.245741 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.245749 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.245757 4777 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8be48c9-46ce-4570-ab17-85835584b238-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.559450 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-8vztl"] Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.569770 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-8vztl"] Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.925346 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-65bb697bdb-4pwlb"] Nov 24 17:21:42 crc kubenswrapper[4777]: E1124 17:21:42.925772 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8be48c9-46ce-4570-ab17-85835584b238" containerName="init" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.925789 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8be48c9-46ce-4570-ab17-85835584b238" containerName="init" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.925981 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8be48c9-46ce-4570-ab17-85835584b238" containerName="init" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.927496 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.930515 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.930674 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.958099 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65bb697bdb-4pwlb"] Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.965106 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27a38d66-1471-470a-95e9-c4bfa5ba415b-logs\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.965157 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27a38d66-1471-470a-95e9-c4bfa5ba415b-config-data-custom\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.965197 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27a38d66-1471-470a-95e9-c4bfa5ba415b-combined-ca-bundle\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.965336 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxqlx\" (UniqueName: \"kubernetes.io/projected/27a38d66-1471-470a-95e9-c4bfa5ba415b-kube-api-access-dxqlx\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.965649 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27a38d66-1471-470a-95e9-c4bfa5ba415b-config-data\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.977584 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a38d66-1471-470a-95e9-c4bfa5ba415b-public-tls-certs\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:42 crc kubenswrapper[4777]: I1124 17:21:42.977668 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a38d66-1471-470a-95e9-c4bfa5ba415b-internal-tls-certs\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.080223 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a38d66-1471-470a-95e9-c4bfa5ba415b-internal-tls-certs\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.080300 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27a38d66-1471-470a-95e9-c4bfa5ba415b-logs\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.080339 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27a38d66-1471-470a-95e9-c4bfa5ba415b-config-data-custom\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.080394 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27a38d66-1471-470a-95e9-c4bfa5ba415b-combined-ca-bundle\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.080456 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxqlx\" (UniqueName: \"kubernetes.io/projected/27a38d66-1471-470a-95e9-c4bfa5ba415b-kube-api-access-dxqlx\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.080589 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27a38d66-1471-470a-95e9-c4bfa5ba415b-config-data\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.080699 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a38d66-1471-470a-95e9-c4bfa5ba415b-public-tls-certs\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.082255 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27a38d66-1471-470a-95e9-c4bfa5ba415b-logs\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.086557 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27a38d66-1471-470a-95e9-c4bfa5ba415b-config-data\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.087128 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a38d66-1471-470a-95e9-c4bfa5ba415b-public-tls-certs\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.087732 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27a38d66-1471-470a-95e9-c4bfa5ba415b-config-data-custom\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.088938 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27a38d66-1471-470a-95e9-c4bfa5ba415b-combined-ca-bundle\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.092765 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a38d66-1471-470a-95e9-c4bfa5ba415b-internal-tls-certs\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.102248 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxqlx\" (UniqueName: \"kubernetes.io/projected/27a38d66-1471-470a-95e9-c4bfa5ba415b-kube-api-access-dxqlx\") pod \"barbican-api-65bb697bdb-4pwlb\" (UID: \"27a38d66-1471-470a-95e9-c4bfa5ba415b\") " pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.205907 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" event={"ID":"33638f0f-d115-472d-8de8-3d459bc6c19c","Type":"ContainerStarted","Data":"8533cbb4cbc2c29d634a0c93caeff0ed22f0bdb3caca2ebb76c693d6aa9aec1e"} Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.207181 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.207210 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.225007 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6wwt4" event={"ID":"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba","Type":"ContainerStarted","Data":"a81bea5b353663e05407a4b03afb1b05194ce88382029dba7dc2d427b5606e5b"} Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.240560 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" podStartSLOduration=4.240539824 podStartE2EDuration="4.240539824s" podCreationTimestamp="2025-11-24 17:21:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:21:43.230618768 +0000 UTC m=+1261.389813817" watchObservedRunningTime="2025-11-24 17:21:43.240539824 +0000 UTC m=+1261.399734873" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.257788 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-6wwt4" podStartSLOduration=4.109634946 podStartE2EDuration="1m4.257768489s" podCreationTimestamp="2025-11-24 17:20:39 +0000 UTC" firstStartedPulling="2025-11-24 17:20:41.709063396 +0000 UTC m=+1199.868258445" lastFinishedPulling="2025-11-24 17:21:41.857196939 +0000 UTC m=+1260.016391988" observedRunningTime="2025-11-24 17:21:43.25641301 +0000 UTC m=+1261.415608059" watchObservedRunningTime="2025-11-24 17:21:43.257768489 +0000 UTC m=+1261.416963538" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.272037 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.281264 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8be48c9-46ce-4570-ab17-85835584b238" path="/var/lib/kubelet/pods/e8be48c9-46ce-4570-ab17-85835584b238/volumes" Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.281762 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-57hm5" event={"ID":"a5ad0d8c-d098-4a37-a8f3-2563fff8b620","Type":"ContainerStarted","Data":"8ec747f5df55dc39b3d1b3cf4cc66787156bf54495e2daaed8ebb86d86f29b99"} Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.281791 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" event={"ID":"b1602c3f-a6a0-40b7-a56b-df49845a3242","Type":"ContainerStarted","Data":"2673ad43798a4047facf2be370033e7ace5b7cceb2017e644a17f0479554fc80"} Nov 24 17:21:43 crc kubenswrapper[4777]: I1124 17:21:43.361548 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" podStartSLOduration=5.361531653 podStartE2EDuration="5.361531653s" podCreationTimestamp="2025-11-24 17:21:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:21:43.356124967 +0000 UTC m=+1261.515320026" watchObservedRunningTime="2025-11-24 17:21:43.361531653 +0000 UTC m=+1261.520726702" Nov 24 17:21:44 crc kubenswrapper[4777]: I1124 17:21:44.269315 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:44 crc kubenswrapper[4777]: I1124 17:21:44.274114 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:21:44 crc kubenswrapper[4777]: I1124 17:21:44.274160 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:21:45 crc kubenswrapper[4777]: I1124 17:21:45.044081 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-sync-57hm5" podStartSLOduration=5.188827417 podStartE2EDuration="1m6.044062141s" podCreationTimestamp="2025-11-24 17:20:39 +0000 UTC" firstStartedPulling="2025-11-24 17:20:41.001151192 +0000 UTC m=+1199.160346241" lastFinishedPulling="2025-11-24 17:21:41.856385916 +0000 UTC m=+1260.015580965" observedRunningTime="2025-11-24 17:21:43.395272433 +0000 UTC m=+1261.554467482" watchObservedRunningTime="2025-11-24 17:21:45.044062141 +0000 UTC m=+1263.203257180" Nov 24 17:21:45 crc kubenswrapper[4777]: I1124 17:21:45.052474 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65bb697bdb-4pwlb"] Nov 24 17:21:45 crc kubenswrapper[4777]: W1124 17:21:45.081171 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27a38d66_1471_470a_95e9_c4bfa5ba415b.slice/crio-2250f568805b9332eae1edc6ebd71f99fedd5e02e9adc55a3da9e112865ec61e WatchSource:0}: Error finding container 2250f568805b9332eae1edc6ebd71f99fedd5e02e9adc55a3da9e112865ec61e: Status 404 returned error can't find the container with id 2250f568805b9332eae1edc6ebd71f99fedd5e02e9adc55a3da9e112865ec61e Nov 24 17:21:45 crc kubenswrapper[4777]: I1124 17:21:45.291342 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65bb697bdb-4pwlb" event={"ID":"27a38d66-1471-470a-95e9-c4bfa5ba415b","Type":"ContainerStarted","Data":"2250f568805b9332eae1edc6ebd71f99fedd5e02e9adc55a3da9e112865ec61e"} Nov 24 17:21:45 crc kubenswrapper[4777]: I1124 17:21:45.296669 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5d57df9989-fbq7p" event={"ID":"dded6598-74bf-4395-88e0-37152b4dd16c","Type":"ContainerStarted","Data":"7dedf71f3169c7c84eb56859d811bc7827d43935a25604d2eab3fa7476f7d680"} Nov 24 17:21:45 crc kubenswrapper[4777]: I1124 17:21:45.306943 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-845875d46-n8hgg" event={"ID":"d173b040-b120-4c77-ba3c-b13c41a6ac37","Type":"ContainerStarted","Data":"ffe4bdd7ce409e30410166a72ef78bfc3c49888a41997820cc396b73a24566e5"} Nov 24 17:21:45 crc kubenswrapper[4777]: I1124 17:21:45.322836 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5d57df9989-fbq7p" podStartSLOduration=2.826002779 podStartE2EDuration="7.322819666s" podCreationTimestamp="2025-11-24 17:21:38 +0000 UTC" firstStartedPulling="2025-11-24 17:21:40.091659424 +0000 UTC m=+1258.250854473" lastFinishedPulling="2025-11-24 17:21:44.588476311 +0000 UTC m=+1262.747671360" observedRunningTime="2025-11-24 17:21:45.317301487 +0000 UTC m=+1263.476496536" watchObservedRunningTime="2025-11-24 17:21:45.322819666 +0000 UTC m=+1263.482014715" Nov 24 17:21:46 crc kubenswrapper[4777]: I1124 17:21:46.318547 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5d57df9989-fbq7p" event={"ID":"dded6598-74bf-4395-88e0-37152b4dd16c","Type":"ContainerStarted","Data":"694df904bfbc338d2176920278cd7b4db23bc92642e419b6c6225f2c6a702931"} Nov 24 17:21:46 crc kubenswrapper[4777]: I1124 17:21:46.321191 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-845875d46-n8hgg" event={"ID":"d173b040-b120-4c77-ba3c-b13c41a6ac37","Type":"ContainerStarted","Data":"503685b7e2184a49f89571dc609b2075f9ec4f7510d0af42615e796ed81b2c75"} Nov 24 17:21:46 crc kubenswrapper[4777]: I1124 17:21:46.323382 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65bb697bdb-4pwlb" event={"ID":"27a38d66-1471-470a-95e9-c4bfa5ba415b","Type":"ContainerStarted","Data":"c34148de9d53fd3c595e7bb00e649dfed12b1c2867b542632d2c5d5238c1f443"} Nov 24 17:21:46 crc kubenswrapper[4777]: I1124 17:21:46.323535 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:46 crc kubenswrapper[4777]: I1124 17:21:46.323604 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65bb697bdb-4pwlb" event={"ID":"27a38d66-1471-470a-95e9-c4bfa5ba415b","Type":"ContainerStarted","Data":"df6d7bf57d05267f475886c9e4dcd3fd1a3ffcf3bec452058e2e23529046572f"} Nov 24 17:21:46 crc kubenswrapper[4777]: I1124 17:21:46.323696 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:46 crc kubenswrapper[4777]: I1124 17:21:46.363518 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-845875d46-n8hgg" podStartSLOduration=3.640977452 podStartE2EDuration="8.363495278s" podCreationTimestamp="2025-11-24 17:21:38 +0000 UTC" firstStartedPulling="2025-11-24 17:21:39.865759379 +0000 UTC m=+1258.024954428" lastFinishedPulling="2025-11-24 17:21:44.588277205 +0000 UTC m=+1262.747472254" observedRunningTime="2025-11-24 17:21:46.346631763 +0000 UTC m=+1264.505826812" watchObservedRunningTime="2025-11-24 17:21:46.363495278 +0000 UTC m=+1264.522690327" Nov 24 17:21:46 crc kubenswrapper[4777]: I1124 17:21:46.427738 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-65bb697bdb-4pwlb" podStartSLOduration=4.427712834 podStartE2EDuration="4.427712834s" podCreationTimestamp="2025-11-24 17:21:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:21:46.367380179 +0000 UTC m=+1264.526575228" watchObservedRunningTime="2025-11-24 17:21:46.427712834 +0000 UTC m=+1264.586907883" Nov 24 17:21:47 crc kubenswrapper[4777]: I1124 17:21:47.416794 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" podUID="33638f0f-d115-472d-8de8-3d459bc6c19c" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 17:21:47 crc kubenswrapper[4777]: I1124 17:21:47.424942 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:48 crc kubenswrapper[4777]: I1124 17:21:48.891312 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:49 crc kubenswrapper[4777]: I1124 17:21:49.440403 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:21:49 crc kubenswrapper[4777]: I1124 17:21:49.535750 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-84jqq"] Nov 24 17:21:49 crc kubenswrapper[4777]: I1124 17:21:49.536068 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" podUID="14a0dfe7-ab97-4cf8-bd83-542d1410e61d" containerName="dnsmasq-dns" containerID="cri-o://ca273ad75ed251795bf946baba0feb47292d34f34d6aba8b1d4f27ebc5f8b270" gracePeriod=10 Nov 24 17:21:50 crc kubenswrapper[4777]: I1124 17:21:50.402823 4777 generic.go:334] "Generic (PLEG): container finished" podID="14a0dfe7-ab97-4cf8-bd83-542d1410e61d" containerID="ca273ad75ed251795bf946baba0feb47292d34f34d6aba8b1d4f27ebc5f8b270" exitCode=0 Nov 24 17:21:50 crc kubenswrapper[4777]: I1124 17:21:50.402897 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" event={"ID":"14a0dfe7-ab97-4cf8-bd83-542d1410e61d","Type":"ContainerDied","Data":"ca273ad75ed251795bf946baba0feb47292d34f34d6aba8b1d4f27ebc5f8b270"} Nov 24 17:21:50 crc kubenswrapper[4777]: I1124 17:21:50.660305 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" podUID="14a0dfe7-ab97-4cf8-bd83-542d1410e61d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.168:5353: connect: connection refused" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.378851 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.434325 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" event={"ID":"14a0dfe7-ab97-4cf8-bd83-542d1410e61d","Type":"ContainerDied","Data":"2f9f67ae8ec25dabd7c2cb789f85709e201e859b802194400f4d1b48bc74d8cd"} Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.434367 4777 scope.go:117] "RemoveContainer" containerID="ca273ad75ed251795bf946baba0feb47292d34f34d6aba8b1d4f27ebc5f8b270" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.434391 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-84jqq" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.470754 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-config\") pod \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.470813 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-ovsdbserver-nb\") pod \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.470909 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8f9s\" (UniqueName: \"kubernetes.io/projected/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-kube-api-access-s8f9s\") pod \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.470941 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-dns-swift-storage-0\") pod \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.471061 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-ovsdbserver-sb\") pod \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.471078 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-dns-svc\") pod \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\" (UID: \"14a0dfe7-ab97-4cf8-bd83-542d1410e61d\") " Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.489275 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-kube-api-access-s8f9s" (OuterVolumeSpecName: "kube-api-access-s8f9s") pod "14a0dfe7-ab97-4cf8-bd83-542d1410e61d" (UID: "14a0dfe7-ab97-4cf8-bd83-542d1410e61d"). InnerVolumeSpecName "kube-api-access-s8f9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.529925 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "14a0dfe7-ab97-4cf8-bd83-542d1410e61d" (UID: "14a0dfe7-ab97-4cf8-bd83-542d1410e61d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.560388 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-config" (OuterVolumeSpecName: "config") pod "14a0dfe7-ab97-4cf8-bd83-542d1410e61d" (UID: "14a0dfe7-ab97-4cf8-bd83-542d1410e61d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.561426 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "14a0dfe7-ab97-4cf8-bd83-542d1410e61d" (UID: "14a0dfe7-ab97-4cf8-bd83-542d1410e61d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.567810 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "14a0dfe7-ab97-4cf8-bd83-542d1410e61d" (UID: "14a0dfe7-ab97-4cf8-bd83-542d1410e61d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.581658 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "14a0dfe7-ab97-4cf8-bd83-542d1410e61d" (UID: "14a0dfe7-ab97-4cf8-bd83-542d1410e61d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.586668 4777 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.586729 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.586740 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.586752 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.586768 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8f9s\" (UniqueName: \"kubernetes.io/projected/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-kube-api-access-s8f9s\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.691367 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a0dfe7-ab97-4cf8-bd83-542d1410e61d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.783615 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-84jqq"] Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.792008 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-84jqq"] Nov 24 17:21:52 crc kubenswrapper[4777]: I1124 17:21:52.984554 4777 scope.go:117] "RemoveContainer" containerID="638b6c66ff3b9de454777656942a7865c7adcf89f0fcdfe31f018ba89175be80" Nov 24 17:21:53 crc kubenswrapper[4777]: I1124 17:21:53.263752 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14a0dfe7-ab97-4cf8-bd83-542d1410e61d" path="/var/lib/kubelet/pods/14a0dfe7-ab97-4cf8-bd83-542d1410e61d/volumes" Nov 24 17:21:53 crc kubenswrapper[4777]: E1124 17:21:53.357332 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"sg-core\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="c0eeb390-6ffc-41c8-a816-0b4454730e07" Nov 24 17:21:53 crc kubenswrapper[4777]: I1124 17:21:53.449343 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0eeb390-6ffc-41c8-a816-0b4454730e07","Type":"ContainerStarted","Data":"06afe8fe5773f3cf973f33c455ae72d9dda6f08c0d265ad6e15cbfc57f4376e5"} Nov 24 17:21:53 crc kubenswrapper[4777]: I1124 17:21:53.449541 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c0eeb390-6ffc-41c8-a816-0b4454730e07" containerName="ceilometer-notification-agent" containerID="cri-o://28cd9ebd6d1385ee8cc3d47a866c0c2c36fbee6fda63ce837a511c472a6be062" gracePeriod=30 Nov 24 17:21:53 crc kubenswrapper[4777]: I1124 17:21:53.449779 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 17:21:53 crc kubenswrapper[4777]: I1124 17:21:53.450065 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c0eeb390-6ffc-41c8-a816-0b4454730e07" containerName="proxy-httpd" containerID="cri-o://06afe8fe5773f3cf973f33c455ae72d9dda6f08c0d265ad6e15cbfc57f4376e5" gracePeriod=30 Nov 24 17:21:53 crc kubenswrapper[4777]: I1124 17:21:53.459587 4777 generic.go:334] "Generic (PLEG): container finished" podID="a5ad0d8c-d098-4a37-a8f3-2563fff8b620" containerID="8ec747f5df55dc39b3d1b3cf4cc66787156bf54495e2daaed8ebb86d86f29b99" exitCode=0 Nov 24 17:21:53 crc kubenswrapper[4777]: I1124 17:21:53.459678 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-57hm5" event={"ID":"a5ad0d8c-d098-4a37-a8f3-2563fff8b620","Type":"ContainerDied","Data":"8ec747f5df55dc39b3d1b3cf4cc66787156bf54495e2daaed8ebb86d86f29b99"} Nov 24 17:21:54 crc kubenswrapper[4777]: I1124 17:21:54.478274 4777 generic.go:334] "Generic (PLEG): container finished" podID="95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba" containerID="a81bea5b353663e05407a4b03afb1b05194ce88382029dba7dc2d427b5606e5b" exitCode=0 Nov 24 17:21:54 crc kubenswrapper[4777]: I1124 17:21:54.478355 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6wwt4" event={"ID":"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba","Type":"ContainerDied","Data":"a81bea5b353663e05407a4b03afb1b05194ce88382029dba7dc2d427b5606e5b"} Nov 24 17:21:54 crc kubenswrapper[4777]: I1124 17:21:54.527562 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:21:54 crc kubenswrapper[4777]: I1124 17:21:54.792802 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:54 crc kubenswrapper[4777]: I1124 17:21:54.840251 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65bb697bdb-4pwlb" Nov 24 17:21:54 crc kubenswrapper[4777]: I1124 17:21:54.920735 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-79b9d8bcdd-vfnwm"] Nov 24 17:21:54 crc kubenswrapper[4777]: I1124 17:21:54.956713 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" podUID="33638f0f-d115-472d-8de8-3d459bc6c19c" containerName="barbican-api" containerID="cri-o://8533cbb4cbc2c29d634a0c93caeff0ed22f0bdb3caca2ebb76c693d6aa9aec1e" gracePeriod=30 Nov 24 17:21:54 crc kubenswrapper[4777]: I1124 17:21:54.959153 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" podUID="33638f0f-d115-472d-8de8-3d459bc6c19c" containerName="barbican-api-log" containerID="cri-o://80a456e5941ed70510ff0ae6327a8bc1cea6d791e29114eeee24f07063cf8411" gracePeriod=30 Nov 24 17:21:54 crc kubenswrapper[4777]: I1124 17:21:54.997468 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.053166 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zszjk\" (UniqueName: \"kubernetes.io/projected/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-kube-api-access-zszjk\") pod \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.053258 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-scripts\") pod \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.053296 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-certs\") pod \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.053348 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-combined-ca-bundle\") pod \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.053384 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-config-data\") pod \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\" (UID: \"a5ad0d8c-d098-4a37-a8f3-2563fff8b620\") " Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.059162 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-kube-api-access-zszjk" (OuterVolumeSpecName: "kube-api-access-zszjk") pod "a5ad0d8c-d098-4a37-a8f3-2563fff8b620" (UID: "a5ad0d8c-d098-4a37-a8f3-2563fff8b620"). InnerVolumeSpecName "kube-api-access-zszjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.061661 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-certs" (OuterVolumeSpecName: "certs") pod "a5ad0d8c-d098-4a37-a8f3-2563fff8b620" (UID: "a5ad0d8c-d098-4a37-a8f3-2563fff8b620"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.064594 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-scripts" (OuterVolumeSpecName: "scripts") pod "a5ad0d8c-d098-4a37-a8f3-2563fff8b620" (UID: "a5ad0d8c-d098-4a37-a8f3-2563fff8b620"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.099549 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-config-data" (OuterVolumeSpecName: "config-data") pod "a5ad0d8c-d098-4a37-a8f3-2563fff8b620" (UID: "a5ad0d8c-d098-4a37-a8f3-2563fff8b620"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.120023 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5ad0d8c-d098-4a37-a8f3-2563fff8b620" (UID: "a5ad0d8c-d098-4a37-a8f3-2563fff8b620"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.155709 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zszjk\" (UniqueName: \"kubernetes.io/projected/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-kube-api-access-zszjk\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.155740 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.155770 4777 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.155781 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.155789 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5ad0d8c-d098-4a37-a8f3-2563fff8b620-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.487508 4777 generic.go:334] "Generic (PLEG): container finished" podID="33638f0f-d115-472d-8de8-3d459bc6c19c" containerID="80a456e5941ed70510ff0ae6327a8bc1cea6d791e29114eeee24f07063cf8411" exitCode=143 Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.487567 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" event={"ID":"33638f0f-d115-472d-8de8-3d459bc6c19c","Type":"ContainerDied","Data":"80a456e5941ed70510ff0ae6327a8bc1cea6d791e29114eeee24f07063cf8411"} Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.490594 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-57hm5" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.491026 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-57hm5" event={"ID":"a5ad0d8c-d098-4a37-a8f3-2563fff8b620","Type":"ContainerDied","Data":"7f36932b14ded8561e413452fd2480b8d210336d357414795bb651cc7d9657f4"} Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.491043 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f36932b14ded8561e413452fd2480b8d210336d357414795bb651cc7d9657f4" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.703120 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-storageinit-mqq7w"] Nov 24 17:21:55 crc kubenswrapper[4777]: E1124 17:21:55.703480 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a0dfe7-ab97-4cf8-bd83-542d1410e61d" containerName="dnsmasq-dns" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.703495 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a0dfe7-ab97-4cf8-bd83-542d1410e61d" containerName="dnsmasq-dns" Nov 24 17:21:55 crc kubenswrapper[4777]: E1124 17:21:55.703510 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5ad0d8c-d098-4a37-a8f3-2563fff8b620" containerName="cloudkitty-db-sync" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.703517 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5ad0d8c-d098-4a37-a8f3-2563fff8b620" containerName="cloudkitty-db-sync" Nov 24 17:21:55 crc kubenswrapper[4777]: E1124 17:21:55.703540 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a0dfe7-ab97-4cf8-bd83-542d1410e61d" containerName="init" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.703546 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a0dfe7-ab97-4cf8-bd83-542d1410e61d" containerName="init" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.703722 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5ad0d8c-d098-4a37-a8f3-2563fff8b620" containerName="cloudkitty-db-sync" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.703735 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="14a0dfe7-ab97-4cf8-bd83-542d1410e61d" containerName="dnsmasq-dns" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.704444 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.706905 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.708505 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.708797 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.708932 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-l2r9w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.709538 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.735584 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-mqq7w"] Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.768251 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkx5t\" (UniqueName: \"kubernetes.io/projected/085c54f5-afa5-431a-84d1-75dad9050d2d-kube-api-access-pkx5t\") pod \"cloudkitty-storageinit-mqq7w\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.768580 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-combined-ca-bundle\") pod \"cloudkitty-storageinit-mqq7w\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.768657 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-scripts\") pod \"cloudkitty-storageinit-mqq7w\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.768717 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-config-data\") pod \"cloudkitty-storageinit-mqq7w\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.768744 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/085c54f5-afa5-431a-84d1-75dad9050d2d-certs\") pod \"cloudkitty-storageinit-mqq7w\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.872169 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-config-data\") pod \"cloudkitty-storageinit-mqq7w\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.872235 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/085c54f5-afa5-431a-84d1-75dad9050d2d-certs\") pod \"cloudkitty-storageinit-mqq7w\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.872279 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkx5t\" (UniqueName: \"kubernetes.io/projected/085c54f5-afa5-431a-84d1-75dad9050d2d-kube-api-access-pkx5t\") pod \"cloudkitty-storageinit-mqq7w\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.872308 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-combined-ca-bundle\") pod \"cloudkitty-storageinit-mqq7w\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.872384 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-scripts\") pod \"cloudkitty-storageinit-mqq7w\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.881032 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-scripts\") pod \"cloudkitty-storageinit-mqq7w\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.886488 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/085c54f5-afa5-431a-84d1-75dad9050d2d-certs\") pod \"cloudkitty-storageinit-mqq7w\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.888955 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-combined-ca-bundle\") pod \"cloudkitty-storageinit-mqq7w\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.893236 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkx5t\" (UniqueName: \"kubernetes.io/projected/085c54f5-afa5-431a-84d1-75dad9050d2d-kube-api-access-pkx5t\") pod \"cloudkitty-storageinit-mqq7w\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:55 crc kubenswrapper[4777]: I1124 17:21:55.906231 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-config-data\") pod \"cloudkitty-storageinit-mqq7w\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.018552 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.185122 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.282624 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-etc-machine-id\") pod \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.282742 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-config-data\") pod \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.282778 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-combined-ca-bundle\") pod \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.282859 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-db-sync-config-data\") pod \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.282954 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l8gn\" (UniqueName: \"kubernetes.io/projected/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-kube-api-access-5l8gn\") pod \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.283032 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-scripts\") pod \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\" (UID: \"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba\") " Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.283829 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba" (UID: "95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.290367 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-scripts" (OuterVolumeSpecName: "scripts") pod "95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba" (UID: "95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.293825 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-kube-api-access-5l8gn" (OuterVolumeSpecName: "kube-api-access-5l8gn") pod "95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba" (UID: "95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba"). InnerVolumeSpecName "kube-api-access-5l8gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.297165 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba" (UID: "95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.337510 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba" (UID: "95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.357400 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-config-data" (OuterVolumeSpecName: "config-data") pod "95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba" (UID: "95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.388759 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.388886 4777 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.388901 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.388913 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.388928 4777 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.388940 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l8gn\" (UniqueName: \"kubernetes.io/projected/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba-kube-api-access-5l8gn\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.502813 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6wwt4" event={"ID":"95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba","Type":"ContainerDied","Data":"27aacf67417c552f473245bc1dd5bfe51f1aaa7fff068370bdcec5f149d5b306"} Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.502864 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27aacf67417c552f473245bc1dd5bfe51f1aaa7fff068370bdcec5f149d5b306" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.502891 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6wwt4" Nov 24 17:21:56 crc kubenswrapper[4777]: W1124 17:21:56.523915 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod085c54f5_afa5_431a_84d1_75dad9050d2d.slice/crio-3cee2fc5d846441ada5eac6680a6ac85a810b65b98966437d5e768b7edeaad57 WatchSource:0}: Error finding container 3cee2fc5d846441ada5eac6680a6ac85a810b65b98966437d5e768b7edeaad57: Status 404 returned error can't find the container with id 3cee2fc5d846441ada5eac6680a6ac85a810b65b98966437d5e768b7edeaad57 Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.527016 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-mqq7w"] Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.695201 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5984748495-2rxn6" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.770113 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-77b7887964-hxn5r"] Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.770315 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-77b7887964-hxn5r" podUID="6885283a-63d3-44a9-928c-39aff5a46e93" containerName="neutron-api" containerID="cri-o://efdb2752b15ceedc4b1933eabf81a88f9b4c0110a61f8e05eb3f4dbe01cd537c" gracePeriod=30 Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.770704 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-77b7887964-hxn5r" podUID="6885283a-63d3-44a9-928c-39aff5a46e93" containerName="neutron-httpd" containerID="cri-o://b49906183d185d99ed7bf30b881efa446b8a6139d8008b6ee5e4218e587e7324" gracePeriod=30 Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.835360 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:21:56 crc kubenswrapper[4777]: E1124 17:21:56.835713 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba" containerName="cinder-db-sync" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.835799 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba" containerName="cinder-db-sync" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.836028 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba" containerName="cinder-db-sync" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.837256 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.843333 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.843390 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.843595 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.843790 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9kzgd" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.868852 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.899022 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghpfc\" (UniqueName: \"kubernetes.io/projected/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-kube-api-access-ghpfc\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.899067 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.899091 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-config-data\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.899205 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.899257 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.899295 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-scripts\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.932213 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-vlhkl"] Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.945354 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:56 crc kubenswrapper[4777]: I1124 17:21:56.952425 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-vlhkl"] Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.005267 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.005328 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-config\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.005358 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-scripts\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.005387 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghpfc\" (UniqueName: \"kubernetes.io/projected/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-kube-api-access-ghpfc\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.005405 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.005425 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-config-data\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.005486 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg2nz\" (UniqueName: \"kubernetes.io/projected/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-kube-api-access-wg2nz\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.005521 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.005540 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-dns-svc\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.005560 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.005581 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.005604 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.007393 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.016038 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-config-data\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.017543 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.017586 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.017941 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-scripts\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.046367 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghpfc\" (UniqueName: \"kubernetes.io/projected/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-kube-api-access-ghpfc\") pod \"cinder-scheduler-0\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " pod="openstack/cinder-scheduler-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.109717 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.110125 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-config\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.110212 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg2nz\" (UniqueName: \"kubernetes.io/projected/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-kube-api-access-wg2nz\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.110238 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.110259 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-dns-svc\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.110280 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.111090 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.111649 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.112027 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-dns-svc\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.112061 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-config\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.112357 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.130143 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.130558 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg2nz\" (UniqueName: \"kubernetes.io/projected/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-kube-api-access-wg2nz\") pod \"dnsmasq-dns-6578955fd5-vlhkl\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.132081 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.135519 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.149685 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.210023 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.211544 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-etc-machine-id\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.211598 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-config-data-custom\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.211678 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-scripts\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.211698 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-config-data\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.211743 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtv7m\" (UniqueName: \"kubernetes.io/projected/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-kube-api-access-qtv7m\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.211773 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.211806 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-logs\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.278400 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.314405 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-config-data-custom\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.314520 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-scripts\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.314540 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-config-data\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.314594 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtv7m\" (UniqueName: \"kubernetes.io/projected/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-kube-api-access-qtv7m\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.314638 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.314671 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-logs\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.314698 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-etc-machine-id\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.314764 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-etc-machine-id\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.335825 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtv7m\" (UniqueName: \"kubernetes.io/projected/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-kube-api-access-qtv7m\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.439332 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-logs\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.442501 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-scripts\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.449430 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-config-data-custom\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.449997 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.450042 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-config-data\") pod \"cinder-api-0\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.548391 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-mqq7w" event={"ID":"085c54f5-afa5-431a-84d1-75dad9050d2d","Type":"ContainerStarted","Data":"fa36391d822c8dd41c854944890ec302f52245eb3aa7a319184bd147967a8c9d"} Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.548709 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-mqq7w" event={"ID":"085c54f5-afa5-431a-84d1-75dad9050d2d","Type":"ContainerStarted","Data":"3cee2fc5d846441ada5eac6680a6ac85a810b65b98966437d5e768b7edeaad57"} Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.559233 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 17:21:57 crc kubenswrapper[4777]: I1124 17:21:57.925263 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:21:58 crc kubenswrapper[4777]: I1124 17:21:58.027724 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-vlhkl"] Nov 24 17:21:58 crc kubenswrapper[4777]: I1124 17:21:58.215492 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:21:58 crc kubenswrapper[4777]: W1124 17:21:58.227046 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48a5b3ec_e2fe_46d6_af20_2b441a9a27bd.slice/crio-3e7623921478a8218fa92f576b3cc5a9882605cb60d8989e60443de476a88571 WatchSource:0}: Error finding container 3e7623921478a8218fa92f576b3cc5a9882605cb60d8989e60443de476a88571: Status 404 returned error can't find the container with id 3e7623921478a8218fa92f576b3cc5a9882605cb60d8989e60443de476a88571 Nov 24 17:21:58 crc kubenswrapper[4777]: I1124 17:21:58.571617 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd","Type":"ContainerStarted","Data":"3e7623921478a8218fa92f576b3cc5a9882605cb60d8989e60443de476a88571"} Nov 24 17:21:58 crc kubenswrapper[4777]: I1124 17:21:58.573559 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc","Type":"ContainerStarted","Data":"0ffa2d46e1643751e5c513767152190814c8817c7f626dbaceabac1c0c0f485a"} Nov 24 17:21:58 crc kubenswrapper[4777]: I1124 17:21:58.578841 4777 generic.go:334] "Generic (PLEG): container finished" podID="6885283a-63d3-44a9-928c-39aff5a46e93" containerID="b49906183d185d99ed7bf30b881efa446b8a6139d8008b6ee5e4218e587e7324" exitCode=0 Nov 24 17:21:58 crc kubenswrapper[4777]: I1124 17:21:58.578897 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77b7887964-hxn5r" event={"ID":"6885283a-63d3-44a9-928c-39aff5a46e93","Type":"ContainerDied","Data":"b49906183d185d99ed7bf30b881efa446b8a6139d8008b6ee5e4218e587e7324"} Nov 24 17:21:58 crc kubenswrapper[4777]: I1124 17:21:58.583268 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" event={"ID":"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b","Type":"ContainerStarted","Data":"613f79a77c81e83f4703c48496ec8d318d1ee5443e95fb13f708299ffbc6ca87"} Nov 24 17:21:58 crc kubenswrapper[4777]: I1124 17:21:58.604171 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-storageinit-mqq7w" podStartSLOduration=3.604155602 podStartE2EDuration="3.604155602s" podCreationTimestamp="2025-11-24 17:21:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:21:58.601521626 +0000 UTC m=+1276.760716675" watchObservedRunningTime="2025-11-24 17:21:58.604155602 +0000 UTC m=+1276.763350651" Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.629298 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd","Type":"ContainerStarted","Data":"5401ef1add4a6a05820876e814f93d33402154c4d4da21203a0b3c72b4b89611"} Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.632344 4777 generic.go:334] "Generic (PLEG): container finished" podID="c0eeb390-6ffc-41c8-a816-0b4454730e07" containerID="28cd9ebd6d1385ee8cc3d47a866c0c2c36fbee6fda63ce837a511c472a6be062" exitCode=0 Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.632400 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0eeb390-6ffc-41c8-a816-0b4454730e07","Type":"ContainerDied","Data":"28cd9ebd6d1385ee8cc3d47a866c0c2c36fbee6fda63ce837a511c472a6be062"} Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.633727 4777 generic.go:334] "Generic (PLEG): container finished" podID="33638f0f-d115-472d-8de8-3d459bc6c19c" containerID="8533cbb4cbc2c29d634a0c93caeff0ed22f0bdb3caca2ebb76c693d6aa9aec1e" exitCode=0 Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.633789 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" event={"ID":"33638f0f-d115-472d-8de8-3d459bc6c19c","Type":"ContainerDied","Data":"8533cbb4cbc2c29d634a0c93caeff0ed22f0bdb3caca2ebb76c693d6aa9aec1e"} Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.633809 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" event={"ID":"33638f0f-d115-472d-8de8-3d459bc6c19c","Type":"ContainerDied","Data":"3b101f47418213fea25f858a695a939559d4f08001662bf6f2046cc56328b4e7"} Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.633819 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b101f47418213fea25f858a695a939559d4f08001662bf6f2046cc56328b4e7" Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.634985 4777 generic.go:334] "Generic (PLEG): container finished" podID="9e3f348f-53d8-49e9-8a18-f9f622c6dc3b" containerID="37c09e91dc2c0e6db87ed920b4774aa23a59477fe2189026536f38a7b786d228" exitCode=0 Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.635007 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" event={"ID":"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b","Type":"ContainerDied","Data":"37c09e91dc2c0e6db87ed920b4774aa23a59477fe2189026536f38a7b786d228"} Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.711773 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.722438 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.779616 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33638f0f-d115-472d-8de8-3d459bc6c19c-logs\") pod \"33638f0f-d115-472d-8de8-3d459bc6c19c\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.779662 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-config-data-custom\") pod \"33638f0f-d115-472d-8de8-3d459bc6c19c\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.779684 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwvfm\" (UniqueName: \"kubernetes.io/projected/33638f0f-d115-472d-8de8-3d459bc6c19c-kube-api-access-pwvfm\") pod \"33638f0f-d115-472d-8de8-3d459bc6c19c\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.779763 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-config-data\") pod \"33638f0f-d115-472d-8de8-3d459bc6c19c\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.779872 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-combined-ca-bundle\") pod \"33638f0f-d115-472d-8de8-3d459bc6c19c\" (UID: \"33638f0f-d115-472d-8de8-3d459bc6c19c\") " Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.784354 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33638f0f-d115-472d-8de8-3d459bc6c19c-logs" (OuterVolumeSpecName: "logs") pod "33638f0f-d115-472d-8de8-3d459bc6c19c" (UID: "33638f0f-d115-472d-8de8-3d459bc6c19c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.788587 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "33638f0f-d115-472d-8de8-3d459bc6c19c" (UID: "33638f0f-d115-472d-8de8-3d459bc6c19c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.795225 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33638f0f-d115-472d-8de8-3d459bc6c19c-kube-api-access-pwvfm" (OuterVolumeSpecName: "kube-api-access-pwvfm") pod "33638f0f-d115-472d-8de8-3d459bc6c19c" (UID: "33638f0f-d115-472d-8de8-3d459bc6c19c"). InnerVolumeSpecName "kube-api-access-pwvfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.883445 4777 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33638f0f-d115-472d-8de8-3d459bc6c19c-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.883489 4777 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.883503 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwvfm\" (UniqueName: \"kubernetes.io/projected/33638f0f-d115-472d-8de8-3d459bc6c19c-kube-api-access-pwvfm\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.890084 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33638f0f-d115-472d-8de8-3d459bc6c19c" (UID: "33638f0f-d115-472d-8de8-3d459bc6c19c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.922835 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-config-data" (OuterVolumeSpecName: "config-data") pod "33638f0f-d115-472d-8de8-3d459bc6c19c" (UID: "33638f0f-d115-472d-8de8-3d459bc6c19c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.996126 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:21:59 crc kubenswrapper[4777]: I1124 17:21:59.996372 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33638f0f-d115-472d-8de8-3d459bc6c19c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.492830 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5877679d68-q5nlj" Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.708231 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd","Type":"ContainerStarted","Data":"a35e4c4bffb7897a92ffc9257c05d0e6aaab089dac5efc537ec10649c3f9d6bb"} Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.708384 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" containerName="cinder-api-log" containerID="cri-o://5401ef1add4a6a05820876e814f93d33402154c4d4da21203a0b3c72b4b89611" gracePeriod=30 Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.708612 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.708646 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" containerName="cinder-api" containerID="cri-o://a35e4c4bffb7897a92ffc9257c05d0e6aaab089dac5efc537ec10649c3f9d6bb" gracePeriod=30 Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.755677 4777 generic.go:334] "Generic (PLEG): container finished" podID="6885283a-63d3-44a9-928c-39aff5a46e93" containerID="efdb2752b15ceedc4b1933eabf81a88f9b4c0110a61f8e05eb3f4dbe01cd537c" exitCode=0 Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.755756 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77b7887964-hxn5r" event={"ID":"6885283a-63d3-44a9-928c-39aff5a46e93","Type":"ContainerDied","Data":"efdb2752b15ceedc4b1933eabf81a88f9b4c0110a61f8e05eb3f4dbe01cd537c"} Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.755781 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77b7887964-hxn5r" event={"ID":"6885283a-63d3-44a9-928c-39aff5a46e93","Type":"ContainerDied","Data":"33b45c448ba260621ea3e3604ef025e97f509e4fce71058422ef02be4aa1ad03"} Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.755793 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33b45c448ba260621ea3e3604ef025e97f509e4fce71058422ef02be4aa1ad03" Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.762131 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.762117499 podStartE2EDuration="3.762117499s" podCreationTimestamp="2025-11-24 17:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:00.758300849 +0000 UTC m=+1278.917495898" watchObservedRunningTime="2025-11-24 17:22:00.762117499 +0000 UTC m=+1278.921312548" Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.801855 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.828316 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.850296 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" event={"ID":"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b","Type":"ContainerStarted","Data":"2101436a7d5cfdd2b38c3f67e716ee9e6e76bfb3f2f11ef6e88577869eee5cd0"} Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.854948 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.916336 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" podStartSLOduration=4.916320483 podStartE2EDuration="4.916320483s" podCreationTimestamp="2025-11-24 17:21:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:00.903878005 +0000 UTC m=+1279.063073054" watchObservedRunningTime="2025-11-24 17:22:00.916320483 +0000 UTC m=+1279.075515532" Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.920825 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-combined-ca-bundle\") pod \"6885283a-63d3-44a9-928c-39aff5a46e93\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.921113 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-ovndb-tls-certs\") pod \"6885283a-63d3-44a9-928c-39aff5a46e93\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.921134 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-httpd-config\") pod \"6885283a-63d3-44a9-928c-39aff5a46e93\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.921170 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmfjx\" (UniqueName: \"kubernetes.io/projected/6885283a-63d3-44a9-928c-39aff5a46e93-kube-api-access-wmfjx\") pod \"6885283a-63d3-44a9-928c-39aff5a46e93\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.921246 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-config\") pod \"6885283a-63d3-44a9-928c-39aff5a46e93\" (UID: \"6885283a-63d3-44a9-928c-39aff5a46e93\") " Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.934262 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-79b9d8bcdd-vfnwm"] Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.945674 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6885283a-63d3-44a9-928c-39aff5a46e93-kube-api-access-wmfjx" (OuterVolumeSpecName: "kube-api-access-wmfjx") pod "6885283a-63d3-44a9-928c-39aff5a46e93" (UID: "6885283a-63d3-44a9-928c-39aff5a46e93"). InnerVolumeSpecName "kube-api-access-wmfjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.947690 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-79b9d8bcdd-vfnwm"] Nov 24 17:22:00 crc kubenswrapper[4777]: I1124 17:22:00.948057 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "6885283a-63d3-44a9-928c-39aff5a46e93" (UID: "6885283a-63d3-44a9-928c-39aff5a46e93"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.023850 4777 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.024957 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmfjx\" (UniqueName: \"kubernetes.io/projected/6885283a-63d3-44a9-928c-39aff5a46e93-kube-api-access-wmfjx\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.093194 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 17:22:01 crc kubenswrapper[4777]: E1124 17:22:01.093665 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6885283a-63d3-44a9-928c-39aff5a46e93" containerName="neutron-api" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.093683 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="6885283a-63d3-44a9-928c-39aff5a46e93" containerName="neutron-api" Nov 24 17:22:01 crc kubenswrapper[4777]: E1124 17:22:01.093698 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6885283a-63d3-44a9-928c-39aff5a46e93" containerName="neutron-httpd" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.093704 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="6885283a-63d3-44a9-928c-39aff5a46e93" containerName="neutron-httpd" Nov 24 17:22:01 crc kubenswrapper[4777]: E1124 17:22:01.093728 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33638f0f-d115-472d-8de8-3d459bc6c19c" containerName="barbican-api-log" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.093734 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="33638f0f-d115-472d-8de8-3d459bc6c19c" containerName="barbican-api-log" Nov 24 17:22:01 crc kubenswrapper[4777]: E1124 17:22:01.093757 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33638f0f-d115-472d-8de8-3d459bc6c19c" containerName="barbican-api" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.093762 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="33638f0f-d115-472d-8de8-3d459bc6c19c" containerName="barbican-api" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.093939 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="6885283a-63d3-44a9-928c-39aff5a46e93" containerName="neutron-httpd" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.093954 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="33638f0f-d115-472d-8de8-3d459bc6c19c" containerName="barbican-api" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.093976 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="6885283a-63d3-44a9-928c-39aff5a46e93" containerName="neutron-api" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.093988 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="33638f0f-d115-472d-8de8-3d459bc6c19c" containerName="barbican-api-log" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.094699 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.100239 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.100415 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.100618 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-7snth" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.122713 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.137291 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6885283a-63d3-44a9-928c-39aff5a46e93" (UID: "6885283a-63d3-44a9-928c-39aff5a46e93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.180090 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-config" (OuterVolumeSpecName: "config") pod "6885283a-63d3-44a9-928c-39aff5a46e93" (UID: "6885283a-63d3-44a9-928c-39aff5a46e93"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.215102 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "6885283a-63d3-44a9-928c-39aff5a46e93" (UID: "6885283a-63d3-44a9-928c-39aff5a46e93"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.237071 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d497520-8510-44d4-b57c-9b8e35ebaa32-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2d497520-8510-44d4-b57c-9b8e35ebaa32\") " pod="openstack/openstackclient" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.237143 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2trdd\" (UniqueName: \"kubernetes.io/projected/2d497520-8510-44d4-b57c-9b8e35ebaa32-kube-api-access-2trdd\") pod \"openstackclient\" (UID: \"2d497520-8510-44d4-b57c-9b8e35ebaa32\") " pod="openstack/openstackclient" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.237233 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d497520-8510-44d4-b57c-9b8e35ebaa32-openstack-config-secret\") pod \"openstackclient\" (UID: \"2d497520-8510-44d4-b57c-9b8e35ebaa32\") " pod="openstack/openstackclient" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.237251 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d497520-8510-44d4-b57c-9b8e35ebaa32-openstack-config\") pod \"openstackclient\" (UID: \"2d497520-8510-44d4-b57c-9b8e35ebaa32\") " pod="openstack/openstackclient" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.237312 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.237328 4777 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.237339 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6885283a-63d3-44a9-928c-39aff5a46e93-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.262253 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33638f0f-d115-472d-8de8-3d459bc6c19c" path="/var/lib/kubelet/pods/33638f0f-d115-472d-8de8-3d459bc6c19c/volumes" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.339400 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d497520-8510-44d4-b57c-9b8e35ebaa32-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2d497520-8510-44d4-b57c-9b8e35ebaa32\") " pod="openstack/openstackclient" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.339475 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2trdd\" (UniqueName: \"kubernetes.io/projected/2d497520-8510-44d4-b57c-9b8e35ebaa32-kube-api-access-2trdd\") pod \"openstackclient\" (UID: \"2d497520-8510-44d4-b57c-9b8e35ebaa32\") " pod="openstack/openstackclient" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.339567 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d497520-8510-44d4-b57c-9b8e35ebaa32-openstack-config-secret\") pod \"openstackclient\" (UID: \"2d497520-8510-44d4-b57c-9b8e35ebaa32\") " pod="openstack/openstackclient" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.339585 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d497520-8510-44d4-b57c-9b8e35ebaa32-openstack-config\") pod \"openstackclient\" (UID: \"2d497520-8510-44d4-b57c-9b8e35ebaa32\") " pod="openstack/openstackclient" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.340560 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d497520-8510-44d4-b57c-9b8e35ebaa32-openstack-config\") pod \"openstackclient\" (UID: \"2d497520-8510-44d4-b57c-9b8e35ebaa32\") " pod="openstack/openstackclient" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.343179 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d497520-8510-44d4-b57c-9b8e35ebaa32-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2d497520-8510-44d4-b57c-9b8e35ebaa32\") " pod="openstack/openstackclient" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.344417 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d497520-8510-44d4-b57c-9b8e35ebaa32-openstack-config-secret\") pod \"openstackclient\" (UID: \"2d497520-8510-44d4-b57c-9b8e35ebaa32\") " pod="openstack/openstackclient" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.360867 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2trdd\" (UniqueName: \"kubernetes.io/projected/2d497520-8510-44d4-b57c-9b8e35ebaa32-kube-api-access-2trdd\") pod \"openstackclient\" (UID: \"2d497520-8510-44d4-b57c-9b8e35ebaa32\") " pod="openstack/openstackclient" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.403422 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.621556 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.628877 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-55f98d7b-5njfr" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.860791 4777 generic.go:334] "Generic (PLEG): container finished" podID="48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" containerID="5401ef1add4a6a05820876e814f93d33402154c4d4da21203a0b3c72b4b89611" exitCode=143 Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.861052 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd","Type":"ContainerDied","Data":"5401ef1add4a6a05820876e814f93d33402154c4d4da21203a0b3c72b4b89611"} Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.882029 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc","Type":"ContainerStarted","Data":"b4c139be37d104936eb8d44ed204802ab5d43d254b70b41636b9bb8cbe6a25ab"} Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.882656 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77b7887964-hxn5r" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.906506 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.230264877 podStartE2EDuration="5.906489184s" podCreationTimestamp="2025-11-24 17:21:56 +0000 UTC" firstStartedPulling="2025-11-24 17:21:57.903601809 +0000 UTC m=+1276.062796858" lastFinishedPulling="2025-11-24 17:21:59.579826116 +0000 UTC m=+1277.739021165" observedRunningTime="2025-11-24 17:22:01.90600198 +0000 UTC m=+1280.065197029" watchObservedRunningTime="2025-11-24 17:22:01.906489184 +0000 UTC m=+1280.065684233" Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.936272 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 17:22:01 crc kubenswrapper[4777]: W1124 17:22:01.949099 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d497520_8510_44d4_b57c_9b8e35ebaa32.slice/crio-9d42fa1494bea66072786ffb2e8aaeada898ad46f28c2d4a54f3df09b5a4e5c9 WatchSource:0}: Error finding container 9d42fa1494bea66072786ffb2e8aaeada898ad46f28c2d4a54f3df09b5a4e5c9: Status 404 returned error can't find the container with id 9d42fa1494bea66072786ffb2e8aaeada898ad46f28c2d4a54f3df09b5a4e5c9 Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.949145 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-77b7887964-hxn5r"] Nov 24 17:22:01 crc kubenswrapper[4777]: I1124 17:22:01.970693 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-77b7887964-hxn5r"] Nov 24 17:22:02 crc kubenswrapper[4777]: I1124 17:22:02.211818 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 17:22:02 crc kubenswrapper[4777]: I1124 17:22:02.905117 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc","Type":"ContainerStarted","Data":"6fae24142d182a6e9d42789ab70b41809f7e6755e1daca9f8123d863209efcf6"} Nov 24 17:22:02 crc kubenswrapper[4777]: I1124 17:22:02.908842 4777 generic.go:334] "Generic (PLEG): container finished" podID="085c54f5-afa5-431a-84d1-75dad9050d2d" containerID="fa36391d822c8dd41c854944890ec302f52245eb3aa7a319184bd147967a8c9d" exitCode=0 Nov 24 17:22:02 crc kubenswrapper[4777]: I1124 17:22:02.908884 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-mqq7w" event={"ID":"085c54f5-afa5-431a-84d1-75dad9050d2d","Type":"ContainerDied","Data":"fa36391d822c8dd41c854944890ec302f52245eb3aa7a319184bd147967a8c9d"} Nov 24 17:22:02 crc kubenswrapper[4777]: I1124 17:22:02.910926 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2d497520-8510-44d4-b57c-9b8e35ebaa32","Type":"ContainerStarted","Data":"9d42fa1494bea66072786ffb2e8aaeada898ad46f28c2d4a54f3df09b5a4e5c9"} Nov 24 17:22:03 crc kubenswrapper[4777]: I1124 17:22:03.261236 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6885283a-63d3-44a9-928c-39aff5a46e93" path="/var/lib/kubelet/pods/6885283a-63d3-44a9-928c-39aff5a46e93/volumes" Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.402753 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.523010 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/085c54f5-afa5-431a-84d1-75dad9050d2d-certs\") pod \"085c54f5-afa5-431a-84d1-75dad9050d2d\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.523140 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-combined-ca-bundle\") pod \"085c54f5-afa5-431a-84d1-75dad9050d2d\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.523171 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-config-data\") pod \"085c54f5-afa5-431a-84d1-75dad9050d2d\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.523243 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkx5t\" (UniqueName: \"kubernetes.io/projected/085c54f5-afa5-431a-84d1-75dad9050d2d-kube-api-access-pkx5t\") pod \"085c54f5-afa5-431a-84d1-75dad9050d2d\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.523325 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-scripts\") pod \"085c54f5-afa5-431a-84d1-75dad9050d2d\" (UID: \"085c54f5-afa5-431a-84d1-75dad9050d2d\") " Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.530041 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-scripts" (OuterVolumeSpecName: "scripts") pod "085c54f5-afa5-431a-84d1-75dad9050d2d" (UID: "085c54f5-afa5-431a-84d1-75dad9050d2d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.531629 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/085c54f5-afa5-431a-84d1-75dad9050d2d-certs" (OuterVolumeSpecName: "certs") pod "085c54f5-afa5-431a-84d1-75dad9050d2d" (UID: "085c54f5-afa5-431a-84d1-75dad9050d2d"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.551620 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/085c54f5-afa5-431a-84d1-75dad9050d2d-kube-api-access-pkx5t" (OuterVolumeSpecName: "kube-api-access-pkx5t") pod "085c54f5-afa5-431a-84d1-75dad9050d2d" (UID: "085c54f5-afa5-431a-84d1-75dad9050d2d"). InnerVolumeSpecName "kube-api-access-pkx5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.558057 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "085c54f5-afa5-431a-84d1-75dad9050d2d" (UID: "085c54f5-afa5-431a-84d1-75dad9050d2d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.562920 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-config-data" (OuterVolumeSpecName: "config-data") pod "085c54f5-afa5-431a-84d1-75dad9050d2d" (UID: "085c54f5-afa5-431a-84d1-75dad9050d2d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.584924 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" podUID="33638f0f-d115-472d-8de8-3d459bc6c19c" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.182:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.584902 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-79b9d8bcdd-vfnwm" podUID="33638f0f-d115-472d-8de8-3d459bc6c19c" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.182:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.625672 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkx5t\" (UniqueName: \"kubernetes.io/projected/085c54f5-afa5-431a-84d1-75dad9050d2d-kube-api-access-pkx5t\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.625709 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.625719 4777 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/085c54f5-afa5-431a-84d1-75dad9050d2d-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.625728 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.625740 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085c54f5-afa5-431a-84d1-75dad9050d2d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.942205 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-mqq7w" event={"ID":"085c54f5-afa5-431a-84d1-75dad9050d2d","Type":"ContainerDied","Data":"3cee2fc5d846441ada5eac6680a6ac85a810b65b98966437d5e768b7edeaad57"} Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.942242 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cee2fc5d846441ada5eac6680a6ac85a810b65b98966437d5e768b7edeaad57" Nov 24 17:22:04 crc kubenswrapper[4777]: I1124 17:22:04.942295 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-mqq7w" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.200613 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 17:22:05 crc kubenswrapper[4777]: E1124 17:22:05.201459 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085c54f5-afa5-431a-84d1-75dad9050d2d" containerName="cloudkitty-storageinit" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.201577 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="085c54f5-afa5-431a-84d1-75dad9050d2d" containerName="cloudkitty-storageinit" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.201926 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="085c54f5-afa5-431a-84d1-75dad9050d2d" containerName="cloudkitty-storageinit" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.202730 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.208104 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.208456 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.208623 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-l2r9w" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.208870 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.213986 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.216260 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.303163 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-vlhkl"] Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.303196 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58bd69657f-bbqzr"] Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.307252 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" podUID="9e3f348f-53d8-49e9-8a18-f9f622c6dc3b" containerName="dnsmasq-dns" containerID="cri-o://2101436a7d5cfdd2b38c3f67e716ee9e6e76bfb3f2f11ef6e88577869eee5cd0" gracePeriod=10 Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.308643 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58bd69657f-bbqzr"] Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.308749 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.309423 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.338352 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.338556 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-scripts\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.338589 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/bded43ac-a1a1-4a4f-a344-177b6befe17c-certs\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.338623 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-config-data\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.338695 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.338730 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhkkp\" (UniqueName: \"kubernetes.io/projected/bded43ac-a1a1-4a4f-a344-177b6befe17c-kube-api-access-hhkkp\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.400294 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.402082 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.415663 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.424328 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.461377 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-scripts\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.461736 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/bded43ac-a1a1-4a4f-a344-177b6befe17c-certs\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.461804 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-config-data\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.461842 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-dns-svc\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.461989 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-config\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.462073 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.462167 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhkkp\" (UniqueName: \"kubernetes.io/projected/bded43ac-a1a1-4a4f-a344-177b6befe17c-kube-api-access-hhkkp\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.462242 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-ovsdbserver-sb\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.462264 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-dns-swift-storage-0\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.462342 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.462412 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-ovsdbserver-nb\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.462450 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b889l\" (UniqueName: \"kubernetes.io/projected/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-kube-api-access-b889l\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.471924 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.483449 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-config-data\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.489394 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.508458 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-scripts\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.521926 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/bded43ac-a1a1-4a4f-a344-177b6befe17c-certs\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.532764 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhkkp\" (UniqueName: \"kubernetes.io/projected/bded43ac-a1a1-4a4f-a344-177b6befe17c-kube-api-access-hhkkp\") pod \"cloudkitty-proc-0\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.564442 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-ovsdbserver-nb\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.564647 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b889l\" (UniqueName: \"kubernetes.io/projected/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-kube-api-access-b889l\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.564829 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.565018 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-dns-svc\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.565201 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/091af110-ca68-49b9-ba84-537dcfc6cb4a-logs\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.565308 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-config\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.565412 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.565514 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkhkf\" (UniqueName: \"kubernetes.io/projected/091af110-ca68-49b9-ba84-537dcfc6cb4a-kube-api-access-xkhkf\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.565589 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/091af110-ca68-49b9-ba84-537dcfc6cb4a-certs\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.565707 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-scripts\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.565803 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-dns-swift-storage-0\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.565882 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-ovsdbserver-sb\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.566055 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-config-data\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.566898 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-config\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.567461 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-dns-svc\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.567674 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-ovsdbserver-sb\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.567876 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-dns-swift-storage-0\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.569882 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-ovsdbserver-nb\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.584600 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b889l\" (UniqueName: \"kubernetes.io/projected/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-kube-api-access-b889l\") pod \"dnsmasq-dns-58bd69657f-bbqzr\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.649119 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.668159 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.668263 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/091af110-ca68-49b9-ba84-537dcfc6cb4a-logs\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.668294 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.668314 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkhkf\" (UniqueName: \"kubernetes.io/projected/091af110-ca68-49b9-ba84-537dcfc6cb4a-kube-api-access-xkhkf\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.668329 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/091af110-ca68-49b9-ba84-537dcfc6cb4a-certs\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.668368 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-scripts\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.668400 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-config-data\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.674303 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/091af110-ca68-49b9-ba84-537dcfc6cb4a-logs\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.676735 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.677710 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.678472 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/091af110-ca68-49b9-ba84-537dcfc6cb4a-certs\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.683659 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-scripts\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.694999 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-config-data\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.726318 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkhkf\" (UniqueName: \"kubernetes.io/projected/091af110-ca68-49b9-ba84-537dcfc6cb4a-kube-api-access-xkhkf\") pod \"cloudkitty-api-0\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.737356 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 17:22:05 crc kubenswrapper[4777]: I1124 17:22:05.826084 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.036160 4777 generic.go:334] "Generic (PLEG): container finished" podID="9e3f348f-53d8-49e9-8a18-f9f622c6dc3b" containerID="2101436a7d5cfdd2b38c3f67e716ee9e6e76bfb3f2f11ef6e88577869eee5cd0" exitCode=0 Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.036202 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" event={"ID":"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b","Type":"ContainerDied","Data":"2101436a7d5cfdd2b38c3f67e716ee9e6e76bfb3f2f11ef6e88577869eee5cd0"} Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.056217 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.192012 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-ovsdbserver-sb\") pod \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.192397 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-ovsdbserver-nb\") pod \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.192459 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-dns-svc\") pod \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.192622 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-dns-swift-storage-0\") pod \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.192669 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-config\") pod \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.192729 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg2nz\" (UniqueName: \"kubernetes.io/projected/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-kube-api-access-wg2nz\") pod \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\" (UID: \"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b\") " Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.204317 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-kube-api-access-wg2nz" (OuterVolumeSpecName: "kube-api-access-wg2nz") pod "9e3f348f-53d8-49e9-8a18-f9f622c6dc3b" (UID: "9e3f348f-53d8-49e9-8a18-f9f622c6dc3b"). InnerVolumeSpecName "kube-api-access-wg2nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.254201 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9e3f348f-53d8-49e9-8a18-f9f622c6dc3b" (UID: "9e3f348f-53d8-49e9-8a18-f9f622c6dc3b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.274103 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9e3f348f-53d8-49e9-8a18-f9f622c6dc3b" (UID: "9e3f348f-53d8-49e9-8a18-f9f622c6dc3b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.277088 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9e3f348f-53d8-49e9-8a18-f9f622c6dc3b" (UID: "9e3f348f-53d8-49e9-8a18-f9f622c6dc3b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.292993 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-config" (OuterVolumeSpecName: "config") pod "9e3f348f-53d8-49e9-8a18-f9f622c6dc3b" (UID: "9e3f348f-53d8-49e9-8a18-f9f622c6dc3b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.294530 4777 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.294562 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.294572 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg2nz\" (UniqueName: \"kubernetes.io/projected/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-kube-api-access-wg2nz\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.294583 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.294592 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.333231 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9e3f348f-53d8-49e9-8a18-f9f622c6dc3b" (UID: "9e3f348f-53d8-49e9-8a18-f9f622c6dc3b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.380410 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58bd69657f-bbqzr"] Nov 24 17:22:06 crc kubenswrapper[4777]: W1124 17:22:06.392586 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2933da9d_6e63_4382_9e7c_6fc6437a7d1c.slice/crio-5878fec6149fb6de39210826d6e51240fe56734dd8ed3f2d44c70adc241a93cb WatchSource:0}: Error finding container 5878fec6149fb6de39210826d6e51240fe56734dd8ed3f2d44c70adc241a93cb: Status 404 returned error can't find the container with id 5878fec6149fb6de39210826d6e51240fe56734dd8ed3f2d44c70adc241a93cb Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.396156 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.619556 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 17:22:06 crc kubenswrapper[4777]: I1124 17:22:06.627624 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.051302 4777 generic.go:334] "Generic (PLEG): container finished" podID="2933da9d-6e63-4382-9e7c-6fc6437a7d1c" containerID="ca52531873febeefc73dc4b0c52b73511302b230f603103fa06605719d3bc668" exitCode=0 Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.051551 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" event={"ID":"2933da9d-6e63-4382-9e7c-6fc6437a7d1c","Type":"ContainerDied","Data":"ca52531873febeefc73dc4b0c52b73511302b230f603103fa06605719d3bc668"} Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.052284 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" event={"ID":"2933da9d-6e63-4382-9e7c-6fc6437a7d1c","Type":"ContainerStarted","Data":"5878fec6149fb6de39210826d6e51240fe56734dd8ed3f2d44c70adc241a93cb"} Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.081592 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"bded43ac-a1a1-4a4f-a344-177b6befe17c","Type":"ContainerStarted","Data":"3d252745a9cf8cf4e73151c59c63801b53671df40759f50b94dbe86c5ccef068"} Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.085080 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"091af110-ca68-49b9-ba84-537dcfc6cb4a","Type":"ContainerStarted","Data":"ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7"} Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.085115 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"091af110-ca68-49b9-ba84-537dcfc6cb4a","Type":"ContainerStarted","Data":"a7938f67ae25caa2dfe56e017e90910fd9a3fd635194ad242627ccbb1655f706"} Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.097536 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" event={"ID":"9e3f348f-53d8-49e9-8a18-f9f622c6dc3b","Type":"ContainerDied","Data":"613f79a77c81e83f4703c48496ec8d318d1ee5443e95fb13f708299ffbc6ca87"} Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.097591 4777 scope.go:117] "RemoveContainer" containerID="2101436a7d5cfdd2b38c3f67e716ee9e6e76bfb3f2f11ef6e88577869eee5cd0" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.097765 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-vlhkl" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.148316 4777 scope.go:117] "RemoveContainer" containerID="37c09e91dc2c0e6db87ed920b4774aa23a59477fe2189026536f38a7b786d228" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.160029 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-vlhkl"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.184040 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-vlhkl"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.280919 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e3f348f-53d8-49e9-8a18-f9f622c6dc3b" path="/var/lib/kubelet/pods/9e3f348f-53d8-49e9-8a18-f9f622c6dc3b/volumes" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.335041 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-7d7z8"] Nov 24 17:22:07 crc kubenswrapper[4777]: E1124 17:22:07.335545 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e3f348f-53d8-49e9-8a18-f9f622c6dc3b" containerName="dnsmasq-dns" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.335560 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e3f348f-53d8-49e9-8a18-f9f622c6dc3b" containerName="dnsmasq-dns" Nov 24 17:22:07 crc kubenswrapper[4777]: E1124 17:22:07.335593 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e3f348f-53d8-49e9-8a18-f9f622c6dc3b" containerName="init" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.335611 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e3f348f-53d8-49e9-8a18-f9f622c6dc3b" containerName="init" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.335852 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e3f348f-53d8-49e9-8a18-f9f622c6dc3b" containerName="dnsmasq-dns" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.336783 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-7d7z8" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.365868 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-7d7z8"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.419231 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2r5r\" (UniqueName: \"kubernetes.io/projected/3a121eb7-f7d9-4d42-9dba-38225f102982-kube-api-access-b2r5r\") pod \"nova-api-db-create-7d7z8\" (UID: \"3a121eb7-f7d9-4d42-9dba-38225f102982\") " pod="openstack/nova-api-db-create-7d7z8" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.419311 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a121eb7-f7d9-4d42-9dba-38225f102982-operator-scripts\") pod \"nova-api-db-create-7d7z8\" (UID: \"3a121eb7-f7d9-4d42-9dba-38225f102982\") " pod="openstack/nova-api-db-create-7d7z8" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.421914 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-xshsz"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.435769 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xshsz" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.444087 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-xshsz"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.453056 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-a9bb-account-create-5d494"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.454545 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a9bb-account-create-5d494" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.459033 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.518386 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-a9bb-account-create-5d494"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.520876 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/051bfdbe-233c-47db-a89f-ff3c89cb342a-operator-scripts\") pod \"nova-cell0-db-create-xshsz\" (UID: \"051bfdbe-233c-47db-a89f-ff3c89cb342a\") " pod="openstack/nova-cell0-db-create-xshsz" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.520986 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2r5r\" (UniqueName: \"kubernetes.io/projected/3a121eb7-f7d9-4d42-9dba-38225f102982-kube-api-access-b2r5r\") pod \"nova-api-db-create-7d7z8\" (UID: \"3a121eb7-f7d9-4d42-9dba-38225f102982\") " pod="openstack/nova-api-db-create-7d7z8" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.521017 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a121eb7-f7d9-4d42-9dba-38225f102982-operator-scripts\") pod \"nova-api-db-create-7d7z8\" (UID: \"3a121eb7-f7d9-4d42-9dba-38225f102982\") " pod="openstack/nova-api-db-create-7d7z8" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.521057 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfxfn\" (UniqueName: \"kubernetes.io/projected/051bfdbe-233c-47db-a89f-ff3c89cb342a-kube-api-access-vfxfn\") pod \"nova-cell0-db-create-xshsz\" (UID: \"051bfdbe-233c-47db-a89f-ff3c89cb342a\") " pod="openstack/nova-cell0-db-create-xshsz" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.521897 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a121eb7-f7d9-4d42-9dba-38225f102982-operator-scripts\") pod \"nova-api-db-create-7d7z8\" (UID: \"3a121eb7-f7d9-4d42-9dba-38225f102982\") " pod="openstack/nova-api-db-create-7d7z8" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.555639 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2r5r\" (UniqueName: \"kubernetes.io/projected/3a121eb7-f7d9-4d42-9dba-38225f102982-kube-api-access-b2r5r\") pod \"nova-api-db-create-7d7z8\" (UID: \"3a121eb7-f7d9-4d42-9dba-38225f102982\") " pod="openstack/nova-api-db-create-7d7z8" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.618269 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-tpnnn"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.619644 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tpnnn" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.623543 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/762d2534-27f5-4323-882d-24f794b2034b-operator-scripts\") pod \"nova-api-a9bb-account-create-5d494\" (UID: \"762d2534-27f5-4323-882d-24f794b2034b\") " pod="openstack/nova-api-a9bb-account-create-5d494" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.623630 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfxfn\" (UniqueName: \"kubernetes.io/projected/051bfdbe-233c-47db-a89f-ff3c89cb342a-kube-api-access-vfxfn\") pod \"nova-cell0-db-create-xshsz\" (UID: \"051bfdbe-233c-47db-a89f-ff3c89cb342a\") " pod="openstack/nova-cell0-db-create-xshsz" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.623686 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjg9w\" (UniqueName: \"kubernetes.io/projected/762d2534-27f5-4323-882d-24f794b2034b-kube-api-access-rjg9w\") pod \"nova-api-a9bb-account-create-5d494\" (UID: \"762d2534-27f5-4323-882d-24f794b2034b\") " pod="openstack/nova-api-a9bb-account-create-5d494" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.623723 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/051bfdbe-233c-47db-a89f-ff3c89cb342a-operator-scripts\") pod \"nova-cell0-db-create-xshsz\" (UID: \"051bfdbe-233c-47db-a89f-ff3c89cb342a\") " pod="openstack/nova-cell0-db-create-xshsz" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.625626 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/051bfdbe-233c-47db-a89f-ff3c89cb342a-operator-scripts\") pod \"nova-cell0-db-create-xshsz\" (UID: \"051bfdbe-233c-47db-a89f-ff3c89cb342a\") " pod="openstack/nova-cell0-db-create-xshsz" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.631171 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.649160 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfxfn\" (UniqueName: \"kubernetes.io/projected/051bfdbe-233c-47db-a89f-ff3c89cb342a-kube-api-access-vfxfn\") pod \"nova-cell0-db-create-xshsz\" (UID: \"051bfdbe-233c-47db-a89f-ff3c89cb342a\") " pod="openstack/nova-cell0-db-create-xshsz" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.655873 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-tpnnn"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.687997 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-7d7z8" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.710070 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-644f75df95-wzmtb"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.713558 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.717142 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.718035 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.722926 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.726394 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjg9w\" (UniqueName: \"kubernetes.io/projected/762d2534-27f5-4323-882d-24f794b2034b-kube-api-access-rjg9w\") pod \"nova-api-a9bb-account-create-5d494\" (UID: \"762d2534-27f5-4323-882d-24f794b2034b\") " pod="openstack/nova-api-a9bb-account-create-5d494" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.739654 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/762d2534-27f5-4323-882d-24f794b2034b-operator-scripts\") pod \"nova-api-a9bb-account-create-5d494\" (UID: \"762d2534-27f5-4323-882d-24f794b2034b\") " pod="openstack/nova-api-a9bb-account-create-5d494" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.739800 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d273d56f-f421-4a2b-8dd5-faaf7603ee79-operator-scripts\") pod \"nova-cell1-db-create-tpnnn\" (UID: \"d273d56f-f421-4a2b-8dd5-faaf7603ee79\") " pod="openstack/nova-cell1-db-create-tpnnn" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.740023 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfncl\" (UniqueName: \"kubernetes.io/projected/d273d56f-f421-4a2b-8dd5-faaf7603ee79-kube-api-access-rfncl\") pod \"nova-cell1-db-create-tpnnn\" (UID: \"d273d56f-f421-4a2b-8dd5-faaf7603ee79\") " pod="openstack/nova-cell1-db-create-tpnnn" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.741035 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/762d2534-27f5-4323-882d-24f794b2034b-operator-scripts\") pod \"nova-api-a9bb-account-create-5d494\" (UID: \"762d2534-27f5-4323-882d-24f794b2034b\") " pod="openstack/nova-api-a9bb-account-create-5d494" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.781772 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjg9w\" (UniqueName: \"kubernetes.io/projected/762d2534-27f5-4323-882d-24f794b2034b-kube-api-access-rjg9w\") pod \"nova-api-a9bb-account-create-5d494\" (UID: \"762d2534-27f5-4323-882d-24f794b2034b\") " pod="openstack/nova-api-a9bb-account-create-5d494" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.781853 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-644f75df95-wzmtb"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.792697 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-18da-account-create-swzmb"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.798485 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-18da-account-create-swzmb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.804742 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.838314 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xshsz" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.843173 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-run-httpd\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.843206 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-public-tls-certs\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.843246 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-internal-tls-certs\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.843279 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-etc-swift\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.843314 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-log-httpd\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.843366 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d273d56f-f421-4a2b-8dd5-faaf7603ee79-operator-scripts\") pod \"nova-cell1-db-create-tpnnn\" (UID: \"d273d56f-f421-4a2b-8dd5-faaf7603ee79\") " pod="openstack/nova-cell1-db-create-tpnnn" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.843387 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-config-data\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.843433 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfncl\" (UniqueName: \"kubernetes.io/projected/d273d56f-f421-4a2b-8dd5-faaf7603ee79-kube-api-access-rfncl\") pod \"nova-cell1-db-create-tpnnn\" (UID: \"d273d56f-f421-4a2b-8dd5-faaf7603ee79\") " pod="openstack/nova-cell1-db-create-tpnnn" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.843470 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btvqg\" (UniqueName: \"kubernetes.io/projected/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-kube-api-access-btvqg\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.843486 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-combined-ca-bundle\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.847302 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d273d56f-f421-4a2b-8dd5-faaf7603ee79-operator-scripts\") pod \"nova-cell1-db-create-tpnnn\" (UID: \"d273d56f-f421-4a2b-8dd5-faaf7603ee79\") " pod="openstack/nova-cell1-db-create-tpnnn" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.853036 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-18da-account-create-swzmb"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.863573 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfncl\" (UniqueName: \"kubernetes.io/projected/d273d56f-f421-4a2b-8dd5-faaf7603ee79-kube-api-access-rfncl\") pod \"nova-cell1-db-create-tpnnn\" (UID: \"d273d56f-f421-4a2b-8dd5-faaf7603ee79\") " pod="openstack/nova-cell1-db-create-tpnnn" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.885912 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.893343 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a9bb-account-create-5d494" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.922017 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-287c-account-create-gnrdg"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.923800 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-287c-account-create-gnrdg" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.930067 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.945315 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-internal-tls-certs\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.945900 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-etc-swift\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.945945 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56-operator-scripts\") pod \"nova-cell0-18da-account-create-swzmb\" (UID: \"2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56\") " pod="openstack/nova-cell0-18da-account-create-swzmb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.946033 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-log-httpd\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.946126 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-config-data\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.946244 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btvqg\" (UniqueName: \"kubernetes.io/projected/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-kube-api-access-btvqg\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.946273 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-combined-ca-bundle\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.946303 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ds5f\" (UniqueName: \"kubernetes.io/projected/2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56-kube-api-access-2ds5f\") pod \"nova-cell0-18da-account-create-swzmb\" (UID: \"2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56\") " pod="openstack/nova-cell0-18da-account-create-swzmb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.946409 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-run-httpd\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.946434 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-public-tls-certs\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.947888 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-log-httpd\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.948496 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tpnnn" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.949319 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-run-httpd\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.951194 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-config-data\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.954085 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-combined-ca-bundle\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.954816 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-etc-swift\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.961666 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-287c-account-create-gnrdg"] Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.962803 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-internal-tls-certs\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.968873 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-public-tls-certs\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:07 crc kubenswrapper[4777]: I1124 17:22:07.981392 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btvqg\" (UniqueName: \"kubernetes.io/projected/22f2176a-dcae-476f-92b5-eb3d19d9bb4f-kube-api-access-btvqg\") pod \"swift-proxy-644f75df95-wzmtb\" (UID: \"22f2176a-dcae-476f-92b5-eb3d19d9bb4f\") " pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.048382 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cbe8193-bfdb-44bb-94af-21372139923c-operator-scripts\") pod \"nova-cell1-287c-account-create-gnrdg\" (UID: \"7cbe8193-bfdb-44bb-94af-21372139923c\") " pod="openstack/nova-cell1-287c-account-create-gnrdg" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.048433 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdc9v\" (UniqueName: \"kubernetes.io/projected/7cbe8193-bfdb-44bb-94af-21372139923c-kube-api-access-bdc9v\") pod \"nova-cell1-287c-account-create-gnrdg\" (UID: \"7cbe8193-bfdb-44bb-94af-21372139923c\") " pod="openstack/nova-cell1-287c-account-create-gnrdg" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.048540 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ds5f\" (UniqueName: \"kubernetes.io/projected/2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56-kube-api-access-2ds5f\") pod \"nova-cell0-18da-account-create-swzmb\" (UID: \"2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56\") " pod="openstack/nova-cell0-18da-account-create-swzmb" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.048633 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56-operator-scripts\") pod \"nova-cell0-18da-account-create-swzmb\" (UID: \"2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56\") " pod="openstack/nova-cell0-18da-account-create-swzmb" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.049356 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56-operator-scripts\") pod \"nova-cell0-18da-account-create-swzmb\" (UID: \"2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56\") " pod="openstack/nova-cell0-18da-account-create-swzmb" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.069507 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ds5f\" (UniqueName: \"kubernetes.io/projected/2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56-kube-api-access-2ds5f\") pod \"nova-cell0-18da-account-create-swzmb\" (UID: \"2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56\") " pod="openstack/nova-cell0-18da-account-create-swzmb" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.115001 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"091af110-ca68-49b9-ba84-537dcfc6cb4a","Type":"ContainerStarted","Data":"d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833"} Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.115225 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.139511 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=3.13949637 podStartE2EDuration="3.13949637s" podCreationTimestamp="2025-11-24 17:22:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:08.135060382 +0000 UTC m=+1286.294255431" watchObservedRunningTime="2025-11-24 17:22:08.13949637 +0000 UTC m=+1286.298691419" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.147790 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" event={"ID":"2933da9d-6e63-4382-9e7c-6fc6437a7d1c","Type":"ContainerStarted","Data":"51e4bec409efb6b7f41777f798933b79c2faf6cf256bb0117bef7083b7b0380e"} Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.147950 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.150342 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cbe8193-bfdb-44bb-94af-21372139923c-operator-scripts\") pod \"nova-cell1-287c-account-create-gnrdg\" (UID: \"7cbe8193-bfdb-44bb-94af-21372139923c\") " pod="openstack/nova-cell1-287c-account-create-gnrdg" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.150383 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdc9v\" (UniqueName: \"kubernetes.io/projected/7cbe8193-bfdb-44bb-94af-21372139923c-kube-api-access-bdc9v\") pod \"nova-cell1-287c-account-create-gnrdg\" (UID: \"7cbe8193-bfdb-44bb-94af-21372139923c\") " pod="openstack/nova-cell1-287c-account-create-gnrdg" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.151649 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cbe8193-bfdb-44bb-94af-21372139923c-operator-scripts\") pod \"nova-cell1-287c-account-create-gnrdg\" (UID: \"7cbe8193-bfdb-44bb-94af-21372139923c\") " pod="openstack/nova-cell1-287c-account-create-gnrdg" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.151782 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" containerName="cinder-scheduler" containerID="cri-o://b4c139be37d104936eb8d44ed204802ab5d43d254b70b41636b9bb8cbe6a25ab" gracePeriod=30 Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.151912 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" containerName="probe" containerID="cri-o://6fae24142d182a6e9d42789ab70b41809f7e6755e1daca9f8123d863209efcf6" gracePeriod=30 Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.175634 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdc9v\" (UniqueName: \"kubernetes.io/projected/7cbe8193-bfdb-44bb-94af-21372139923c-kube-api-access-bdc9v\") pod \"nova-cell1-287c-account-create-gnrdg\" (UID: \"7cbe8193-bfdb-44bb-94af-21372139923c\") " pod="openstack/nova-cell1-287c-account-create-gnrdg" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.180373 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" podStartSLOduration=3.180356905 podStartE2EDuration="3.180356905s" podCreationTimestamp="2025-11-24 17:22:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:08.164473738 +0000 UTC m=+1286.323668787" watchObservedRunningTime="2025-11-24 17:22:08.180356905 +0000 UTC m=+1286.339551954" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.182439 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.203957 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-18da-account-create-swzmb" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.257757 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-287c-account-create-gnrdg" Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.308709 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.409817 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-7d7z8"] Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.602029 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-xshsz"] Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.700959 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-a9bb-account-create-5d494"] Nov 24 17:22:08 crc kubenswrapper[4777]: I1124 17:22:08.710268 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-tpnnn"] Nov 24 17:22:09 crc kubenswrapper[4777]: I1124 17:22:09.061065 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-644f75df95-wzmtb"] Nov 24 17:22:09 crc kubenswrapper[4777]: I1124 17:22:09.171121 4777 generic.go:334] "Generic (PLEG): container finished" podID="3a121eb7-f7d9-4d42-9dba-38225f102982" containerID="e8ecd53f8e21a9ac85fadb81b7f325d9fab0bc5ac6db8d3bb92bf494b87208f4" exitCode=0 Nov 24 17:22:09 crc kubenswrapper[4777]: I1124 17:22:09.171181 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-7d7z8" event={"ID":"3a121eb7-f7d9-4d42-9dba-38225f102982","Type":"ContainerDied","Data":"e8ecd53f8e21a9ac85fadb81b7f325d9fab0bc5ac6db8d3bb92bf494b87208f4"} Nov 24 17:22:09 crc kubenswrapper[4777]: I1124 17:22:09.171536 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-7d7z8" event={"ID":"3a121eb7-f7d9-4d42-9dba-38225f102982","Type":"ContainerStarted","Data":"7122c4013879f3cdc5f3eebb11881baee22ab7ee2e5b79101d0b5ed742354be7"} Nov 24 17:22:09 crc kubenswrapper[4777]: I1124 17:22:09.178924 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-18da-account-create-swzmb"] Nov 24 17:22:09 crc kubenswrapper[4777]: I1124 17:22:09.180912 4777 generic.go:334] "Generic (PLEG): container finished" podID="7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" containerID="6fae24142d182a6e9d42789ab70b41809f7e6755e1daca9f8123d863209efcf6" exitCode=0 Nov 24 17:22:09 crc kubenswrapper[4777]: I1124 17:22:09.180982 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc","Type":"ContainerDied","Data":"6fae24142d182a6e9d42789ab70b41809f7e6755e1daca9f8123d863209efcf6"} Nov 24 17:22:09 crc kubenswrapper[4777]: I1124 17:22:09.184421 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xshsz" event={"ID":"051bfdbe-233c-47db-a89f-ff3c89cb342a","Type":"ContainerStarted","Data":"a2039e0c12ac7a10a2d79874e41a4a29d716cae02125a7f0711fd29997d465ea"} Nov 24 17:22:09 crc kubenswrapper[4777]: I1124 17:22:09.196606 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-287c-account-create-gnrdg"] Nov 24 17:22:09 crc kubenswrapper[4777]: W1124 17:22:09.581702 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7cbe8193_bfdb_44bb_94af_21372139923c.slice/crio-1d6298776215d080ee876f823624cd0810dcc42949022c48bf4e6b976d6a74e0 WatchSource:0}: Error finding container 1d6298776215d080ee876f823624cd0810dcc42949022c48bf4e6b976d6a74e0: Status 404 returned error can't find the container with id 1d6298776215d080ee876f823624cd0810dcc42949022c48bf4e6b976d6a74e0 Nov 24 17:22:09 crc kubenswrapper[4777]: W1124 17:22:09.597143 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22f2176a_dcae_476f_92b5_eb3d19d9bb4f.slice/crio-82aff28838ec8c25f9287c2bca36d8cfc5919d421a6eb3f223127daddac62d65 WatchSource:0}: Error finding container 82aff28838ec8c25f9287c2bca36d8cfc5919d421a6eb3f223127daddac62d65: Status 404 returned error can't find the container with id 82aff28838ec8c25f9287c2bca36d8cfc5919d421a6eb3f223127daddac62d65 Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.122634 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="c0eeb390-6ffc-41c8-a816-0b4454730e07" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.218246 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tpnnn" event={"ID":"d273d56f-f421-4a2b-8dd5-faaf7603ee79","Type":"ContainerStarted","Data":"955033450ae4da4d2e8d7c8e6e68d31e7b2d1cf85c8de5723855f4d63c00ce7b"} Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.218298 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tpnnn" event={"ID":"d273d56f-f421-4a2b-8dd5-faaf7603ee79","Type":"ContainerStarted","Data":"7583e1a33c81a08f021b2ba940d4c507e41b4b11e5888d94d67e827695bda147"} Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.231369 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-18da-account-create-swzmb" event={"ID":"2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56","Type":"ContainerStarted","Data":"0d7e170b7359e74d361ac51d3a801fc417c2eab27f388ebb016cbfe0ae7b35ef"} Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.246626 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xshsz" event={"ID":"051bfdbe-233c-47db-a89f-ff3c89cb342a","Type":"ContainerStarted","Data":"5baef7c26317fe983dd13937a633921b0e460e0cba388302662a30cfd101fdfd"} Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.250123 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-tpnnn" podStartSLOduration=3.250089746 podStartE2EDuration="3.250089746s" podCreationTimestamp="2025-11-24 17:22:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:10.23702109 +0000 UTC m=+1288.396216159" watchObservedRunningTime="2025-11-24 17:22:10.250089746 +0000 UTC m=+1288.409284795" Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.254795 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-644f75df95-wzmtb" event={"ID":"22f2176a-dcae-476f-92b5-eb3d19d9bb4f","Type":"ContainerStarted","Data":"82aff28838ec8c25f9287c2bca36d8cfc5919d421a6eb3f223127daddac62d65"} Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.263251 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a9bb-account-create-5d494" event={"ID":"762d2534-27f5-4323-882d-24f794b2034b","Type":"ContainerStarted","Data":"4f009c3fbaa25f1413fc6b2a70e9e1eb560a5cfe85b5bec54104bdd1f4012aca"} Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.263298 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a9bb-account-create-5d494" event={"ID":"762d2534-27f5-4323-882d-24f794b2034b","Type":"ContainerStarted","Data":"3a861bf483dde852fcc50eebac90612cb65d01a2de7a625dc8047ca02cbcfb56"} Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.264695 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-xshsz" podStartSLOduration=3.264678226 podStartE2EDuration="3.264678226s" podCreationTimestamp="2025-11-24 17:22:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:10.263308846 +0000 UTC m=+1288.422503895" watchObservedRunningTime="2025-11-24 17:22:10.264678226 +0000 UTC m=+1288.423873275" Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.272634 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="091af110-ca68-49b9-ba84-537dcfc6cb4a" containerName="cloudkitty-api-log" containerID="cri-o://ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7" gracePeriod=30 Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.272876 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="091af110-ca68-49b9-ba84-537dcfc6cb4a" containerName="cloudkitty-api" containerID="cri-o://d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833" gracePeriod=30 Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.272915 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-287c-account-create-gnrdg" event={"ID":"7cbe8193-bfdb-44bb-94af-21372139923c","Type":"ContainerStarted","Data":"0841bed172a9a98f66e2c8b89e20550ea7d92eca0caf45afeb2ab6f81479ab19"} Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.272954 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-287c-account-create-gnrdg" event={"ID":"7cbe8193-bfdb-44bb-94af-21372139923c","Type":"ContainerStarted","Data":"1d6298776215d080ee876f823624cd0810dcc42949022c48bf4e6b976d6a74e0"} Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.290935 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-a9bb-account-create-5d494" podStartSLOduration=3.29092084 podStartE2EDuration="3.29092084s" podCreationTimestamp="2025-11-24 17:22:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:10.278705209 +0000 UTC m=+1288.437900258" watchObservedRunningTime="2025-11-24 17:22:10.29092084 +0000 UTC m=+1288.450115889" Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.306500 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-287c-account-create-gnrdg" podStartSLOduration=3.306482768 podStartE2EDuration="3.306482768s" podCreationTimestamp="2025-11-24 17:22:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:10.301839294 +0000 UTC m=+1288.461034343" watchObservedRunningTime="2025-11-24 17:22:10.306482768 +0000 UTC m=+1288.465677817" Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.774224 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-7d7z8" Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.926433 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a121eb7-f7d9-4d42-9dba-38225f102982-operator-scripts\") pod \"3a121eb7-f7d9-4d42-9dba-38225f102982\" (UID: \"3a121eb7-f7d9-4d42-9dba-38225f102982\") " Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.926574 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2r5r\" (UniqueName: \"kubernetes.io/projected/3a121eb7-f7d9-4d42-9dba-38225f102982-kube-api-access-b2r5r\") pod \"3a121eb7-f7d9-4d42-9dba-38225f102982\" (UID: \"3a121eb7-f7d9-4d42-9dba-38225f102982\") " Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.927599 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a121eb7-f7d9-4d42-9dba-38225f102982-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3a121eb7-f7d9-4d42-9dba-38225f102982" (UID: "3a121eb7-f7d9-4d42-9dba-38225f102982"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:10 crc kubenswrapper[4777]: I1124 17:22:10.945374 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a121eb7-f7d9-4d42-9dba-38225f102982-kube-api-access-b2r5r" (OuterVolumeSpecName: "kube-api-access-b2r5r") pod "3a121eb7-f7d9-4d42-9dba-38225f102982" (UID: "3a121eb7-f7d9-4d42-9dba-38225f102982"). InnerVolumeSpecName "kube-api-access-b2r5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.008795 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.031351 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a121eb7-f7d9-4d42-9dba-38225f102982-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.031384 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2r5r\" (UniqueName: \"kubernetes.io/projected/3a121eb7-f7d9-4d42-9dba-38225f102982-kube-api-access-b2r5r\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.132384 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-scripts\") pod \"091af110-ca68-49b9-ba84-537dcfc6cb4a\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.132479 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-config-data\") pod \"091af110-ca68-49b9-ba84-537dcfc6cb4a\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.132518 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-combined-ca-bundle\") pod \"091af110-ca68-49b9-ba84-537dcfc6cb4a\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.132628 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/091af110-ca68-49b9-ba84-537dcfc6cb4a-logs\") pod \"091af110-ca68-49b9-ba84-537dcfc6cb4a\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.132676 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/091af110-ca68-49b9-ba84-537dcfc6cb4a-certs\") pod \"091af110-ca68-49b9-ba84-537dcfc6cb4a\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.132703 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkhkf\" (UniqueName: \"kubernetes.io/projected/091af110-ca68-49b9-ba84-537dcfc6cb4a-kube-api-access-xkhkf\") pod \"091af110-ca68-49b9-ba84-537dcfc6cb4a\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.132734 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-config-data-custom\") pod \"091af110-ca68-49b9-ba84-537dcfc6cb4a\" (UID: \"091af110-ca68-49b9-ba84-537dcfc6cb4a\") " Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.133237 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/091af110-ca68-49b9-ba84-537dcfc6cb4a-logs" (OuterVolumeSpecName: "logs") pod "091af110-ca68-49b9-ba84-537dcfc6cb4a" (UID: "091af110-ca68-49b9-ba84-537dcfc6cb4a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.137608 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "091af110-ca68-49b9-ba84-537dcfc6cb4a" (UID: "091af110-ca68-49b9-ba84-537dcfc6cb4a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.139958 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-scripts" (OuterVolumeSpecName: "scripts") pod "091af110-ca68-49b9-ba84-537dcfc6cb4a" (UID: "091af110-ca68-49b9-ba84-537dcfc6cb4a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.140709 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/091af110-ca68-49b9-ba84-537dcfc6cb4a-kube-api-access-xkhkf" (OuterVolumeSpecName: "kube-api-access-xkhkf") pod "091af110-ca68-49b9-ba84-537dcfc6cb4a" (UID: "091af110-ca68-49b9-ba84-537dcfc6cb4a"). InnerVolumeSpecName "kube-api-access-xkhkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.146570 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/091af110-ca68-49b9-ba84-537dcfc6cb4a-certs" (OuterVolumeSpecName: "certs") pod "091af110-ca68-49b9-ba84-537dcfc6cb4a" (UID: "091af110-ca68-49b9-ba84-537dcfc6cb4a"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.178847 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.184033 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-config-data" (OuterVolumeSpecName: "config-data") pod "091af110-ca68-49b9-ba84-537dcfc6cb4a" (UID: "091af110-ca68-49b9-ba84-537dcfc6cb4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.194372 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "091af110-ca68-49b9-ba84-537dcfc6cb4a" (UID: "091af110-ca68-49b9-ba84-537dcfc6cb4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.235114 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.235141 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.235152 4777 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/091af110-ca68-49b9-ba84-537dcfc6cb4a-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.235160 4777 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/091af110-ca68-49b9-ba84-537dcfc6cb4a-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.235170 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkhkf\" (UniqueName: \"kubernetes.io/projected/091af110-ca68-49b9-ba84-537dcfc6cb4a-kube-api-access-xkhkf\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.235179 4777 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.235186 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/091af110-ca68-49b9-ba84-537dcfc6cb4a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.302875 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-644f75df95-wzmtb" event={"ID":"22f2176a-dcae-476f-92b5-eb3d19d9bb4f","Type":"ContainerStarted","Data":"27be789659d1f9e57a4fbe3592f8dbc747554784e3a559b8f447ca8fc3ddcfc5"} Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.302928 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.302940 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-644f75df95-wzmtb" event={"ID":"22f2176a-dcae-476f-92b5-eb3d19d9bb4f","Type":"ContainerStarted","Data":"2f0d72a2eba5cf53c9d2f9bacee1bf89120eaeee132487161aae41409795002b"} Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.303020 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.310310 4777 generic.go:334] "Generic (PLEG): container finished" podID="7cbe8193-bfdb-44bb-94af-21372139923c" containerID="0841bed172a9a98f66e2c8b89e20550ea7d92eca0caf45afeb2ab6f81479ab19" exitCode=0 Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.310387 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-287c-account-create-gnrdg" event={"ID":"7cbe8193-bfdb-44bb-94af-21372139923c","Type":"ContainerDied","Data":"0841bed172a9a98f66e2c8b89e20550ea7d92eca0caf45afeb2ab6f81479ab19"} Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.314073 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"bded43ac-a1a1-4a4f-a344-177b6befe17c","Type":"ContainerStarted","Data":"ab8ead3a65d02607b933ff4423b9e4f656d6acc38d7f27b064cdbd7f49ca7d55"} Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.319036 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-7d7z8" event={"ID":"3a121eb7-f7d9-4d42-9dba-38225f102982","Type":"ContainerDied","Data":"7122c4013879f3cdc5f3eebb11881baee22ab7ee2e5b79101d0b5ed742354be7"} Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.319094 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7122c4013879f3cdc5f3eebb11881baee22ab7ee2e5b79101d0b5ed742354be7" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.319143 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-7d7z8" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.332188 4777 generic.go:334] "Generic (PLEG): container finished" podID="091af110-ca68-49b9-ba84-537dcfc6cb4a" containerID="d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833" exitCode=0 Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.332212 4777 generic.go:334] "Generic (PLEG): container finished" podID="091af110-ca68-49b9-ba84-537dcfc6cb4a" containerID="ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7" exitCode=143 Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.332252 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"091af110-ca68-49b9-ba84-537dcfc6cb4a","Type":"ContainerDied","Data":"d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833"} Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.332278 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"091af110-ca68-49b9-ba84-537dcfc6cb4a","Type":"ContainerDied","Data":"ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7"} Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.332289 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"091af110-ca68-49b9-ba84-537dcfc6cb4a","Type":"ContainerDied","Data":"a7938f67ae25caa2dfe56e017e90910fd9a3fd635194ad242627ccbb1655f706"} Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.332303 4777 scope.go:117] "RemoveContainer" containerID="d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.332419 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.342824 4777 generic.go:334] "Generic (PLEG): container finished" podID="051bfdbe-233c-47db-a89f-ff3c89cb342a" containerID="5baef7c26317fe983dd13937a633921b0e460e0cba388302662a30cfd101fdfd" exitCode=0 Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.342897 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xshsz" event={"ID":"051bfdbe-233c-47db-a89f-ff3c89cb342a","Type":"ContainerDied","Data":"5baef7c26317fe983dd13937a633921b0e460e0cba388302662a30cfd101fdfd"} Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.345088 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-644f75df95-wzmtb" podStartSLOduration=4.345067689 podStartE2EDuration="4.345067689s" podCreationTimestamp="2025-11-24 17:22:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:11.328332878 +0000 UTC m=+1289.487527927" watchObservedRunningTime="2025-11-24 17:22:11.345067689 +0000 UTC m=+1289.504262738" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.357540 4777 generic.go:334] "Generic (PLEG): container finished" podID="762d2534-27f5-4323-882d-24f794b2034b" containerID="4f009c3fbaa25f1413fc6b2a70e9e1eb560a5cfe85b5bec54104bdd1f4012aca" exitCode=0 Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.357631 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a9bb-account-create-5d494" event={"ID":"762d2534-27f5-4323-882d-24f794b2034b","Type":"ContainerDied","Data":"4f009c3fbaa25f1413fc6b2a70e9e1eb560a5cfe85b5bec54104bdd1f4012aca"} Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.363662 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=3.160472673 podStartE2EDuration="6.363643703s" podCreationTimestamp="2025-11-24 17:22:05 +0000 UTC" firstStartedPulling="2025-11-24 17:22:06.639472981 +0000 UTC m=+1284.798668030" lastFinishedPulling="2025-11-24 17:22:09.842644011 +0000 UTC m=+1288.001839060" observedRunningTime="2025-11-24 17:22:11.357398884 +0000 UTC m=+1289.516593953" watchObservedRunningTime="2025-11-24 17:22:11.363643703 +0000 UTC m=+1289.522838752" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.370945 4777 generic.go:334] "Generic (PLEG): container finished" podID="d273d56f-f421-4a2b-8dd5-faaf7603ee79" containerID="955033450ae4da4d2e8d7c8e6e68d31e7b2d1cf85c8de5723855f4d63c00ce7b" exitCode=0 Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.371048 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tpnnn" event={"ID":"d273d56f-f421-4a2b-8dd5-faaf7603ee79","Type":"ContainerDied","Data":"955033450ae4da4d2e8d7c8e6e68d31e7b2d1cf85c8de5723855f4d63c00ce7b"} Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.383840 4777 scope.go:117] "RemoveContainer" containerID="ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.386221 4777 generic.go:334] "Generic (PLEG): container finished" podID="2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56" containerID="24e8514d456e8b25c8cf1b4c6dab4f462f6a8a89c1f1f6a5d588c27cef0bbfd4" exitCode=0 Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.386259 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-18da-account-create-swzmb" event={"ID":"2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56","Type":"ContainerDied","Data":"24e8514d456e8b25c8cf1b4c6dab4f462f6a8a89c1f1f6a5d588c27cef0bbfd4"} Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.405368 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.416428 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.424996 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.435418 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 17:22:11 crc kubenswrapper[4777]: E1124 17:22:11.435861 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a121eb7-f7d9-4d42-9dba-38225f102982" containerName="mariadb-database-create" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.435877 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a121eb7-f7d9-4d42-9dba-38225f102982" containerName="mariadb-database-create" Nov 24 17:22:11 crc kubenswrapper[4777]: E1124 17:22:11.435905 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091af110-ca68-49b9-ba84-537dcfc6cb4a" containerName="cloudkitty-api-log" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.435913 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="091af110-ca68-49b9-ba84-537dcfc6cb4a" containerName="cloudkitty-api-log" Nov 24 17:22:11 crc kubenswrapper[4777]: E1124 17:22:11.435938 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091af110-ca68-49b9-ba84-537dcfc6cb4a" containerName="cloudkitty-api" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.435944 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="091af110-ca68-49b9-ba84-537dcfc6cb4a" containerName="cloudkitty-api" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.436258 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="091af110-ca68-49b9-ba84-537dcfc6cb4a" containerName="cloudkitty-api" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.436280 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a121eb7-f7d9-4d42-9dba-38225f102982" containerName="mariadb-database-create" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.436288 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="091af110-ca68-49b9-ba84-537dcfc6cb4a" containerName="cloudkitty-api-log" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.437491 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.444096 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-internal-svc" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.444301 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.444405 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-public-svc" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.450171 4777 scope.go:117] "RemoveContainer" containerID="d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833" Nov 24 17:22:11 crc kubenswrapper[4777]: E1124 17:22:11.451400 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833\": container with ID starting with d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833 not found: ID does not exist" containerID="d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.451438 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833"} err="failed to get container status \"d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833\": rpc error: code = NotFound desc = could not find container \"d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833\": container with ID starting with d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833 not found: ID does not exist" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.451466 4777 scope.go:117] "RemoveContainer" containerID="ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7" Nov 24 17:22:11 crc kubenswrapper[4777]: E1124 17:22:11.452530 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7\": container with ID starting with ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7 not found: ID does not exist" containerID="ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.452576 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7"} err="failed to get container status \"ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7\": rpc error: code = NotFound desc = could not find container \"ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7\": container with ID starting with ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7 not found: ID does not exist" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.452616 4777 scope.go:117] "RemoveContainer" containerID="d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.456930 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833"} err="failed to get container status \"d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833\": rpc error: code = NotFound desc = could not find container \"d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833\": container with ID starting with d19e122ed15b9e7ac28b86ffa890ba1f5b00bf1581c08cf6dbe0a296362bc833 not found: ID does not exist" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.456998 4777 scope.go:117] "RemoveContainer" containerID="ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.462730 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7"} err="failed to get container status \"ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7\": rpc error: code = NotFound desc = could not find container \"ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7\": container with ID starting with ef9010bf923a81df27743aff7c2f6802f2d934e2795d004acfc08657a3ea74a7 not found: ID does not exist" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.474047 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.541354 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.541430 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-scripts\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.541459 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94ad549a-1f18-4ad2-8d56-844e83d35aec-logs\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.541475 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.541500 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.541547 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.541573 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-config-data\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.541597 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/94ad549a-1f18-4ad2-8d56-844e83d35aec-certs\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.541618 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwqlx\" (UniqueName: \"kubernetes.io/projected/94ad549a-1f18-4ad2-8d56-844e83d35aec-kube-api-access-bwqlx\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.642905 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-scripts\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.643220 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94ad549a-1f18-4ad2-8d56-844e83d35aec-logs\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.643240 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.643270 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.643316 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.643341 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-config-data\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.643833 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94ad549a-1f18-4ad2-8d56-844e83d35aec-logs\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.643883 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/94ad549a-1f18-4ad2-8d56-844e83d35aec-certs\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.643912 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwqlx\" (UniqueName: \"kubernetes.io/projected/94ad549a-1f18-4ad2-8d56-844e83d35aec-kube-api-access-bwqlx\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.644056 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.649079 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-scripts\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.649397 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.652069 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-config-data\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.652240 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.658318 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.664776 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.666413 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/94ad549a-1f18-4ad2-8d56-844e83d35aec-certs\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.669716 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwqlx\" (UniqueName: \"kubernetes.io/projected/94ad549a-1f18-4ad2-8d56-844e83d35aec-kube-api-access-bwqlx\") pod \"cloudkitty-api-0\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " pod="openstack/cloudkitty-api-0" Nov 24 17:22:11 crc kubenswrapper[4777]: I1124 17:22:11.769132 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 17:22:13 crc kubenswrapper[4777]: I1124 17:22:13.275450 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="091af110-ca68-49b9-ba84-537dcfc6cb4a" path="/var/lib/kubelet/pods/091af110-ca68-49b9-ba84-537dcfc6cb4a/volumes" Nov 24 17:22:13 crc kubenswrapper[4777]: I1124 17:22:13.466477 4777 generic.go:334] "Generic (PLEG): container finished" podID="7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" containerID="b4c139be37d104936eb8d44ed204802ab5d43d254b70b41636b9bb8cbe6a25ab" exitCode=0 Nov 24 17:22:13 crc kubenswrapper[4777]: I1124 17:22:13.466633 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="bded43ac-a1a1-4a4f-a344-177b6befe17c" containerName="cloudkitty-proc" containerID="cri-o://ab8ead3a65d02607b933ff4423b9e4f656d6acc38d7f27b064cdbd7f49ca7d55" gracePeriod=30 Nov 24 17:22:13 crc kubenswrapper[4777]: I1124 17:22:13.466924 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc","Type":"ContainerDied","Data":"b4c139be37d104936eb8d44ed204802ab5d43d254b70b41636b9bb8cbe6a25ab"} Nov 24 17:22:14 crc kubenswrapper[4777]: I1124 17:22:14.270388 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:22:14 crc kubenswrapper[4777]: I1124 17:22:14.270621 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:22:15 crc kubenswrapper[4777]: I1124 17:22:15.492955 4777 generic.go:334] "Generic (PLEG): container finished" podID="bded43ac-a1a1-4a4f-a344-177b6befe17c" containerID="ab8ead3a65d02607b933ff4423b9e4f656d6acc38d7f27b064cdbd7f49ca7d55" exitCode=0 Nov 24 17:22:15 crc kubenswrapper[4777]: I1124 17:22:15.493013 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"bded43ac-a1a1-4a4f-a344-177b6befe17c","Type":"ContainerDied","Data":"ab8ead3a65d02607b933ff4423b9e4f656d6acc38d7f27b064cdbd7f49ca7d55"} Nov 24 17:22:15 crc kubenswrapper[4777]: I1124 17:22:15.651100 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:22:15 crc kubenswrapper[4777]: I1124 17:22:15.708050 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-hngmc"] Nov 24 17:22:15 crc kubenswrapper[4777]: I1124 17:22:15.708283 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" podUID="b1602c3f-a6a0-40b7-a56b-df49845a3242" containerName="dnsmasq-dns" containerID="cri-o://2673ad43798a4047facf2be370033e7ace5b7cceb2017e644a17f0479554fc80" gracePeriod=10 Nov 24 17:22:16 crc kubenswrapper[4777]: I1124 17:22:16.507127 4777 generic.go:334] "Generic (PLEG): container finished" podID="b1602c3f-a6a0-40b7-a56b-df49845a3242" containerID="2673ad43798a4047facf2be370033e7ace5b7cceb2017e644a17f0479554fc80" exitCode=0 Nov 24 17:22:16 crc kubenswrapper[4777]: I1124 17:22:16.507233 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" event={"ID":"b1602c3f-a6a0-40b7-a56b-df49845a3242","Type":"ContainerDied","Data":"2673ad43798a4047facf2be370033e7ace5b7cceb2017e644a17f0479554fc80"} Nov 24 17:22:18 crc kubenswrapper[4777]: I1124 17:22:18.191243 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:18 crc kubenswrapper[4777]: I1124 17:22:18.191621 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-644f75df95-wzmtb" Nov 24 17:22:19 crc kubenswrapper[4777]: I1124 17:22:19.437826 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" podUID="b1602c3f-a6a0-40b7-a56b-df49845a3242" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.181:5353: connect: connection refused" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.485677 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-287c-account-create-gnrdg" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.486252 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-18da-account-create-swzmb" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.534316 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a9bb-account-create-5d494" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.556043 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ds5f\" (UniqueName: \"kubernetes.io/projected/2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56-kube-api-access-2ds5f\") pod \"2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56\" (UID: \"2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56\") " Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.556278 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56-operator-scripts\") pod \"2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56\" (UID: \"2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56\") " Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.556522 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cbe8193-bfdb-44bb-94af-21372139923c-operator-scripts\") pod \"7cbe8193-bfdb-44bb-94af-21372139923c\" (UID: \"7cbe8193-bfdb-44bb-94af-21372139923c\") " Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.556628 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdc9v\" (UniqueName: \"kubernetes.io/projected/7cbe8193-bfdb-44bb-94af-21372139923c-kube-api-access-bdc9v\") pod \"7cbe8193-bfdb-44bb-94af-21372139923c\" (UID: \"7cbe8193-bfdb-44bb-94af-21372139923c\") " Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.556927 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56" (UID: "2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.558890 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.559618 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cbe8193-bfdb-44bb-94af-21372139923c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7cbe8193-bfdb-44bb-94af-21372139923c" (UID: "7cbe8193-bfdb-44bb-94af-21372139923c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.566183 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cbe8193-bfdb-44bb-94af-21372139923c-kube-api-access-bdc9v" (OuterVolumeSpecName: "kube-api-access-bdc9v") pod "7cbe8193-bfdb-44bb-94af-21372139923c" (UID: "7cbe8193-bfdb-44bb-94af-21372139923c"). InnerVolumeSpecName "kube-api-access-bdc9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.568040 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56-kube-api-access-2ds5f" (OuterVolumeSpecName: "kube-api-access-2ds5f") pod "2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56" (UID: "2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56"). InnerVolumeSpecName "kube-api-access-2ds5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.580629 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-18da-account-create-swzmb" event={"ID":"2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56","Type":"ContainerDied","Data":"0d7e170b7359e74d361ac51d3a801fc417c2eab27f388ebb016cbfe0ae7b35ef"} Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.580837 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d7e170b7359e74d361ac51d3a801fc417c2eab27f388ebb016cbfe0ae7b35ef" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.580943 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-18da-account-create-swzmb" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.598268 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xshsz" event={"ID":"051bfdbe-233c-47db-a89f-ff3c89cb342a","Type":"ContainerDied","Data":"a2039e0c12ac7a10a2d79874e41a4a29d716cae02125a7f0711fd29997d465ea"} Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.598308 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2039e0c12ac7a10a2d79874e41a4a29d716cae02125a7f0711fd29997d465ea" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.610417 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a9bb-account-create-5d494" event={"ID":"762d2534-27f5-4323-882d-24f794b2034b","Type":"ContainerDied","Data":"3a861bf483dde852fcc50eebac90612cb65d01a2de7a625dc8047ca02cbcfb56"} Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.610617 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a861bf483dde852fcc50eebac90612cb65d01a2de7a625dc8047ca02cbcfb56" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.610723 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a9bb-account-create-5d494" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.614778 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-287c-account-create-gnrdg" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.614962 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-287c-account-create-gnrdg" event={"ID":"7cbe8193-bfdb-44bb-94af-21372139923c","Type":"ContainerDied","Data":"1d6298776215d080ee876f823624cd0810dcc42949022c48bf4e6b976d6a74e0"} Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.615007 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d6298776215d080ee876f823624cd0810dcc42949022c48bf4e6b976d6a74e0" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.626753 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tpnnn" event={"ID":"d273d56f-f421-4a2b-8dd5-faaf7603ee79","Type":"ContainerDied","Data":"7583e1a33c81a08f021b2ba940d4c507e41b4b11e5888d94d67e827695bda147"} Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.626803 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7583e1a33c81a08f021b2ba940d4c507e41b4b11e5888d94d67e827695bda147" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.655660 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tpnnn" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.659574 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjg9w\" (UniqueName: \"kubernetes.io/projected/762d2534-27f5-4323-882d-24f794b2034b-kube-api-access-rjg9w\") pod \"762d2534-27f5-4323-882d-24f794b2034b\" (UID: \"762d2534-27f5-4323-882d-24f794b2034b\") " Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.659746 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/762d2534-27f5-4323-882d-24f794b2034b-operator-scripts\") pod \"762d2534-27f5-4323-882d-24f794b2034b\" (UID: \"762d2534-27f5-4323-882d-24f794b2034b\") " Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.660230 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cbe8193-bfdb-44bb-94af-21372139923c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.660249 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdc9v\" (UniqueName: \"kubernetes.io/projected/7cbe8193-bfdb-44bb-94af-21372139923c-kube-api-access-bdc9v\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.660260 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ds5f\" (UniqueName: \"kubernetes.io/projected/2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56-kube-api-access-2ds5f\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.660747 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/762d2534-27f5-4323-882d-24f794b2034b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "762d2534-27f5-4323-882d-24f794b2034b" (UID: "762d2534-27f5-4323-882d-24f794b2034b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.671719 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/762d2534-27f5-4323-882d-24f794b2034b-kube-api-access-rjg9w" (OuterVolumeSpecName: "kube-api-access-rjg9w") pod "762d2534-27f5-4323-882d-24f794b2034b" (UID: "762d2534-27f5-4323-882d-24f794b2034b"). InnerVolumeSpecName "kube-api-access-rjg9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.712060 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xshsz" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.761692 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/051bfdbe-233c-47db-a89f-ff3c89cb342a-operator-scripts\") pod \"051bfdbe-233c-47db-a89f-ff3c89cb342a\" (UID: \"051bfdbe-233c-47db-a89f-ff3c89cb342a\") " Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.761851 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfxfn\" (UniqueName: \"kubernetes.io/projected/051bfdbe-233c-47db-a89f-ff3c89cb342a-kube-api-access-vfxfn\") pod \"051bfdbe-233c-47db-a89f-ff3c89cb342a\" (UID: \"051bfdbe-233c-47db-a89f-ff3c89cb342a\") " Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.761989 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d273d56f-f421-4a2b-8dd5-faaf7603ee79-operator-scripts\") pod \"d273d56f-f421-4a2b-8dd5-faaf7603ee79\" (UID: \"d273d56f-f421-4a2b-8dd5-faaf7603ee79\") " Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.762059 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfncl\" (UniqueName: \"kubernetes.io/projected/d273d56f-f421-4a2b-8dd5-faaf7603ee79-kube-api-access-rfncl\") pod \"d273d56f-f421-4a2b-8dd5-faaf7603ee79\" (UID: \"d273d56f-f421-4a2b-8dd5-faaf7603ee79\") " Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.762619 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjg9w\" (UniqueName: \"kubernetes.io/projected/762d2534-27f5-4323-882d-24f794b2034b-kube-api-access-rjg9w\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.762642 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/762d2534-27f5-4323-882d-24f794b2034b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.764269 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/051bfdbe-233c-47db-a89f-ff3c89cb342a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "051bfdbe-233c-47db-a89f-ff3c89cb342a" (UID: "051bfdbe-233c-47db-a89f-ff3c89cb342a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.764448 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d273d56f-f421-4a2b-8dd5-faaf7603ee79-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d273d56f-f421-4a2b-8dd5-faaf7603ee79" (UID: "d273d56f-f421-4a2b-8dd5-faaf7603ee79"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.768249 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d273d56f-f421-4a2b-8dd5-faaf7603ee79-kube-api-access-rfncl" (OuterVolumeSpecName: "kube-api-access-rfncl") pod "d273d56f-f421-4a2b-8dd5-faaf7603ee79" (UID: "d273d56f-f421-4a2b-8dd5-faaf7603ee79"). InnerVolumeSpecName "kube-api-access-rfncl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.769742 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/051bfdbe-233c-47db-a89f-ff3c89cb342a-kube-api-access-vfxfn" (OuterVolumeSpecName: "kube-api-access-vfxfn") pod "051bfdbe-233c-47db-a89f-ff3c89cb342a" (UID: "051bfdbe-233c-47db-a89f-ff3c89cb342a"). InnerVolumeSpecName "kube-api-access-vfxfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.865093 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfxfn\" (UniqueName: \"kubernetes.io/projected/051bfdbe-233c-47db-a89f-ff3c89cb342a-kube-api-access-vfxfn\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.865172 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d273d56f-f421-4a2b-8dd5-faaf7603ee79-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.865187 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfncl\" (UniqueName: \"kubernetes.io/projected/d273d56f-f421-4a2b-8dd5-faaf7603ee79-kube-api-access-rfncl\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:20 crc kubenswrapper[4777]: I1124 17:22:20.865201 4777 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/051bfdbe-233c-47db-a89f-ff3c89cb342a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.025372 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.051004 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.073081 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-config-data-custom\") pod \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.073183 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-etc-machine-id\") pod \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.073218 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghpfc\" (UniqueName: \"kubernetes.io/projected/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-kube-api-access-ghpfc\") pod \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.073629 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-scripts\") pod \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.073679 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-combined-ca-bundle\") pod \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.073699 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-config-data\") pod \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\" (UID: \"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.076685 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" (UID: "7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.079464 4777 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.080801 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-kube-api-access-ghpfc" (OuterVolumeSpecName: "kube-api-access-ghpfc") pod "7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" (UID: "7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc"). InnerVolumeSpecName "kube-api-access-ghpfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.084762 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" (UID: "7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.098733 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-scripts" (OuterVolumeSpecName: "scripts") pod "7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" (UID: "7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.153164 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.170542 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" (UID: "7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.180470 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-scripts\") pod \"bded43ac-a1a1-4a4f-a344-177b6befe17c\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.180529 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/bded43ac-a1a1-4a4f-a344-177b6befe17c-certs\") pod \"bded43ac-a1a1-4a4f-a344-177b6befe17c\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.180572 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-config-data\") pod \"bded43ac-a1a1-4a4f-a344-177b6befe17c\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.180671 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-combined-ca-bundle\") pod \"bded43ac-a1a1-4a4f-a344-177b6befe17c\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.180735 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-config-data-custom\") pod \"bded43ac-a1a1-4a4f-a344-177b6befe17c\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.180792 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhkkp\" (UniqueName: \"kubernetes.io/projected/bded43ac-a1a1-4a4f-a344-177b6befe17c-kube-api-access-hhkkp\") pod \"bded43ac-a1a1-4a4f-a344-177b6befe17c\" (UID: \"bded43ac-a1a1-4a4f-a344-177b6befe17c\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.181262 4777 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.181282 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghpfc\" (UniqueName: \"kubernetes.io/projected/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-kube-api-access-ghpfc\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.181294 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.181303 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.183592 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-scripts" (OuterVolumeSpecName: "scripts") pod "bded43ac-a1a1-4a4f-a344-177b6befe17c" (UID: "bded43ac-a1a1-4a4f-a344-177b6befe17c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.188237 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bded43ac-a1a1-4a4f-a344-177b6befe17c-certs" (OuterVolumeSpecName: "certs") pod "bded43ac-a1a1-4a4f-a344-177b6befe17c" (UID: "bded43ac-a1a1-4a4f-a344-177b6befe17c"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.193633 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bded43ac-a1a1-4a4f-a344-177b6befe17c-kube-api-access-hhkkp" (OuterVolumeSpecName: "kube-api-access-hhkkp") pod "bded43ac-a1a1-4a4f-a344-177b6befe17c" (UID: "bded43ac-a1a1-4a4f-a344-177b6befe17c"). InnerVolumeSpecName "kube-api-access-hhkkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.201408 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bded43ac-a1a1-4a4f-a344-177b6befe17c" (UID: "bded43ac-a1a1-4a4f-a344-177b6befe17c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.228190 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-config-data" (OuterVolumeSpecName: "config-data") pod "bded43ac-a1a1-4a4f-a344-177b6befe17c" (UID: "bded43ac-a1a1-4a4f-a344-177b6befe17c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.233562 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bded43ac-a1a1-4a4f-a344-177b6befe17c" (UID: "bded43ac-a1a1-4a4f-a344-177b6befe17c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.247977 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-config-data" (OuterVolumeSpecName: "config-data") pod "7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" (UID: "7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.283660 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.283946 4777 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/bded43ac-a1a1-4a4f-a344-177b6befe17c-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.284028 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.284092 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.284164 4777 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bded43ac-a1a1-4a4f-a344-177b6befe17c-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.284228 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.284286 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhkkp\" (UniqueName: \"kubernetes.io/projected/bded43ac-a1a1-4a4f-a344-177b6befe17c-kube-api-access-hhkkp\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.408925 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.487718 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-config\") pod \"b1602c3f-a6a0-40b7-a56b-df49845a3242\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.487771 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-dns-svc\") pod \"b1602c3f-a6a0-40b7-a56b-df49845a3242\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.487804 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-ovsdbserver-sb\") pod \"b1602c3f-a6a0-40b7-a56b-df49845a3242\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.487845 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z54wf\" (UniqueName: \"kubernetes.io/projected/b1602c3f-a6a0-40b7-a56b-df49845a3242-kube-api-access-z54wf\") pod \"b1602c3f-a6a0-40b7-a56b-df49845a3242\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.487985 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-ovsdbserver-nb\") pod \"b1602c3f-a6a0-40b7-a56b-df49845a3242\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.488112 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-dns-swift-storage-0\") pod \"b1602c3f-a6a0-40b7-a56b-df49845a3242\" (UID: \"b1602c3f-a6a0-40b7-a56b-df49845a3242\") " Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.503417 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1602c3f-a6a0-40b7-a56b-df49845a3242-kube-api-access-z54wf" (OuterVolumeSpecName: "kube-api-access-z54wf") pod "b1602c3f-a6a0-40b7-a56b-df49845a3242" (UID: "b1602c3f-a6a0-40b7-a56b-df49845a3242"). InnerVolumeSpecName "kube-api-access-z54wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.577095 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b1602c3f-a6a0-40b7-a56b-df49845a3242" (UID: "b1602c3f-a6a0-40b7-a56b-df49845a3242"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.578851 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-config" (OuterVolumeSpecName: "config") pod "b1602c3f-a6a0-40b7-a56b-df49845a3242" (UID: "b1602c3f-a6a0-40b7-a56b-df49845a3242"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.585092 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b1602c3f-a6a0-40b7-a56b-df49845a3242" (UID: "b1602c3f-a6a0-40b7-a56b-df49845a3242"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.596679 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.608181 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.608250 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z54wf\" (UniqueName: \"kubernetes.io/projected/b1602c3f-a6a0-40b7-a56b-df49845a3242-kube-api-access-z54wf\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.608267 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.600311 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b1602c3f-a6a0-40b7-a56b-df49845a3242" (UID: "b1602c3f-a6a0-40b7-a56b-df49845a3242"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.611177 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b1602c3f-a6a0-40b7-a56b-df49845a3242" (UID: "b1602c3f-a6a0-40b7-a56b-df49845a3242"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.656510 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc","Type":"ContainerDied","Data":"0ffa2d46e1643751e5c513767152190814c8817c7f626dbaceabac1c0c0f485a"} Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.656568 4777 scope.go:117] "RemoveContainer" containerID="6fae24142d182a6e9d42789ab70b41809f7e6755e1daca9f8123d863209efcf6" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.656697 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.660118 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"94ad549a-1f18-4ad2-8d56-844e83d35aec","Type":"ContainerStarted","Data":"1280b04734fef908aac2b889626d0963282ca59afe1eb90ea1366d3c28b44536"} Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.660167 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"94ad549a-1f18-4ad2-8d56-844e83d35aec","Type":"ContainerStarted","Data":"e053e3cda387b7ba7b7e17346f4ee0b28e592e30d06f6edf3f759e1387a710c3"} Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.666501 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" event={"ID":"b1602c3f-a6a0-40b7-a56b-df49845a3242","Type":"ContainerDied","Data":"4c7d48553db4b7f018904050faf1381e28c0bd4fe080df9b3f46e7d7b3f170b4"} Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.666558 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-hngmc" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.668204 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2d497520-8510-44d4-b57c-9b8e35ebaa32","Type":"ContainerStarted","Data":"32026e59a58132916c59eb7dd2b62209277ffc5aaa1d4a65897a6199d4f7c2ac"} Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.680278 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xshsz" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.680370 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"bded43ac-a1a1-4a4f-a344-177b6befe17c","Type":"ContainerDied","Data":"3d252745a9cf8cf4e73151c59c63801b53671df40759f50b94dbe86c5ccef068"} Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.680456 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tpnnn" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.680518 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.709454 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.158091588 podStartE2EDuration="20.70943569s" podCreationTimestamp="2025-11-24 17:22:01 +0000 UTC" firstStartedPulling="2025-11-24 17:22:01.961726792 +0000 UTC m=+1280.120921841" lastFinishedPulling="2025-11-24 17:22:20.513070894 +0000 UTC m=+1298.672265943" observedRunningTime="2025-11-24 17:22:21.694818747 +0000 UTC m=+1299.854013796" watchObservedRunningTime="2025-11-24 17:22:21.70943569 +0000 UTC m=+1299.868630739" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.710427 4777 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.710475 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1602c3f-a6a0-40b7-a56b-df49845a3242-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.714933 4777 scope.go:117] "RemoveContainer" containerID="b4c139be37d104936eb8d44ed204802ab5d43d254b70b41636b9bb8cbe6a25ab" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.750087 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.761684 4777 scope.go:117] "RemoveContainer" containerID="2673ad43798a4047facf2be370033e7ace5b7cceb2017e644a17f0479554fc80" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.770082 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.817180 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:22:21 crc kubenswrapper[4777]: E1124 17:22:21.817583 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="051bfdbe-233c-47db-a89f-ff3c89cb342a" containerName="mariadb-database-create" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.817599 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="051bfdbe-233c-47db-a89f-ff3c89cb342a" containerName="mariadb-database-create" Nov 24 17:22:21 crc kubenswrapper[4777]: E1124 17:22:21.817624 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="762d2534-27f5-4323-882d-24f794b2034b" containerName="mariadb-account-create" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.817632 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="762d2534-27f5-4323-882d-24f794b2034b" containerName="mariadb-account-create" Nov 24 17:22:21 crc kubenswrapper[4777]: E1124 17:22:21.817642 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56" containerName="mariadb-account-create" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.817648 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56" containerName="mariadb-account-create" Nov 24 17:22:21 crc kubenswrapper[4777]: E1124 17:22:21.817660 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bded43ac-a1a1-4a4f-a344-177b6befe17c" containerName="cloudkitty-proc" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.817667 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="bded43ac-a1a1-4a4f-a344-177b6befe17c" containerName="cloudkitty-proc" Nov 24 17:22:21 crc kubenswrapper[4777]: E1124 17:22:21.817688 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" containerName="probe" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.817694 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" containerName="probe" Nov 24 17:22:21 crc kubenswrapper[4777]: E1124 17:22:21.817702 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cbe8193-bfdb-44bb-94af-21372139923c" containerName="mariadb-account-create" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.817707 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cbe8193-bfdb-44bb-94af-21372139923c" containerName="mariadb-account-create" Nov 24 17:22:21 crc kubenswrapper[4777]: E1124 17:22:21.817717 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1602c3f-a6a0-40b7-a56b-df49845a3242" containerName="dnsmasq-dns" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.817724 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1602c3f-a6a0-40b7-a56b-df49845a3242" containerName="dnsmasq-dns" Nov 24 17:22:21 crc kubenswrapper[4777]: E1124 17:22:21.817732 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" containerName="cinder-scheduler" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.817739 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" containerName="cinder-scheduler" Nov 24 17:22:21 crc kubenswrapper[4777]: E1124 17:22:21.817752 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1602c3f-a6a0-40b7-a56b-df49845a3242" containerName="init" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.817757 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1602c3f-a6a0-40b7-a56b-df49845a3242" containerName="init" Nov 24 17:22:21 crc kubenswrapper[4777]: E1124 17:22:21.817773 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d273d56f-f421-4a2b-8dd5-faaf7603ee79" containerName="mariadb-database-create" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.817779 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="d273d56f-f421-4a2b-8dd5-faaf7603ee79" containerName="mariadb-database-create" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.817943 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1602c3f-a6a0-40b7-a56b-df49845a3242" containerName="dnsmasq-dns" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.817980 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" containerName="cinder-scheduler" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.817992 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="051bfdbe-233c-47db-a89f-ff3c89cb342a" containerName="mariadb-database-create" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.818000 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" containerName="probe" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.818008 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="762d2534-27f5-4323-882d-24f794b2034b" containerName="mariadb-account-create" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.818021 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="bded43ac-a1a1-4a4f-a344-177b6befe17c" containerName="cloudkitty-proc" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.818029 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56" containerName="mariadb-account-create" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.818038 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cbe8193-bfdb-44bb-94af-21372139923c" containerName="mariadb-account-create" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.818047 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="d273d56f-f421-4a2b-8dd5-faaf7603ee79" containerName="mariadb-database-create" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.819074 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.830023 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-hngmc"] Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.831386 4777 scope.go:117] "RemoveContainer" containerID="dae56e2bd4581eae9f6c3e1176902d8795dcc89ece8c2d14f46d371fc437f64a" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.832024 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.848249 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-hngmc"] Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.858688 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.868884 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.888519 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.897514 4777 scope.go:117] "RemoveContainer" containerID="ab8ead3a65d02607b933ff4423b9e4f656d6acc38d7f27b064cdbd7f49ca7d55" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.897662 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.899067 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.905287 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.905285 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.933066 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.933137 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqtbr\" (UniqueName: \"kubernetes.io/projected/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-kube-api-access-kqtbr\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.933171 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.933192 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-scripts\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.933208 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-config-data\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:21 crc kubenswrapper[4777]: I1124 17:22:21.933248 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.036091 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-scripts\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.036161 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.036324 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.036423 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.036434 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vmhp\" (UniqueName: \"kubernetes.io/projected/2aa8a135-f57b-4d6b-9444-4fbdfc542476-kube-api-access-5vmhp\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.036475 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-config-data\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.036522 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/2aa8a135-f57b-4d6b-9444-4fbdfc542476-certs\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.036541 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.036562 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqtbr\" (UniqueName: \"kubernetes.io/projected/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-kube-api-access-kqtbr\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.036589 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.036637 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-scripts\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.036654 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-config-data\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.036674 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.041528 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-scripts\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.041558 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.041941 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.042585 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-config-data\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.055545 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqtbr\" (UniqueName: \"kubernetes.io/projected/f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13-kube-api-access-kqtbr\") pod \"cinder-scheduler-0\" (UID: \"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13\") " pod="openstack/cinder-scheduler-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.138108 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/2aa8a135-f57b-4d6b-9444-4fbdfc542476-certs\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.138159 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.138280 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-scripts\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.138310 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.138399 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vmhp\" (UniqueName: \"kubernetes.io/projected/2aa8a135-f57b-4d6b-9444-4fbdfc542476-kube-api-access-5vmhp\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.138423 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-config-data\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.141683 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.142097 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-config-data\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.142454 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/2aa8a135-f57b-4d6b-9444-4fbdfc542476-certs\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.142596 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.144875 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-scripts\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.156798 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vmhp\" (UniqueName: \"kubernetes.io/projected/2aa8a135-f57b-4d6b-9444-4fbdfc542476-kube-api-access-5vmhp\") pod \"cloudkitty-proc-0\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.164444 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.226249 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.653781 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.695346 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"94ad549a-1f18-4ad2-8d56-844e83d35aec","Type":"ContainerStarted","Data":"167e60e62b59ed489e4a5c43f0ae11f7a45c8d96c0ba0017e044a46720f1c382"} Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.696032 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.696947 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13","Type":"ContainerStarted","Data":"ca746590f3b1aa2d2724c24800967a5fadc0775fb2a0742d48ec3881ade15c9e"} Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.721057 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=11.721037244 podStartE2EDuration="11.721037244s" podCreationTimestamp="2025-11-24 17:22:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:22.711639302 +0000 UTC m=+1300.870834351" watchObservedRunningTime="2025-11-24 17:22:22.721037244 +0000 UTC m=+1300.880232293" Nov 24 17:22:22 crc kubenswrapper[4777]: I1124 17:22:22.787109 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.130634 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5rcr4"] Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.132203 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.137026 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.137620 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-npnv4" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.137734 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.159706 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5rcr4"] Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.260946 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-5rcr4\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.261006 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-config-data\") pod \"nova-cell0-conductor-db-sync-5rcr4\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.261044 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md6xw\" (UniqueName: \"kubernetes.io/projected/abf0902b-bd70-49df-966b-4cabf47f75df-kube-api-access-md6xw\") pod \"nova-cell0-conductor-db-sync-5rcr4\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.261077 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-scripts\") pod \"nova-cell0-conductor-db-sync-5rcr4\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.265123 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc" path="/var/lib/kubelet/pods/7f6edd7f-2893-4b0f-bfee-0a8a6f0b75cc/volumes" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.265896 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1602c3f-a6a0-40b7-a56b-df49845a3242" path="/var/lib/kubelet/pods/b1602c3f-a6a0-40b7-a56b-df49845a3242/volumes" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.266589 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bded43ac-a1a1-4a4f-a344-177b6befe17c" path="/var/lib/kubelet/pods/bded43ac-a1a1-4a4f-a344-177b6befe17c/volumes" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.369309 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md6xw\" (UniqueName: \"kubernetes.io/projected/abf0902b-bd70-49df-966b-4cabf47f75df-kube-api-access-md6xw\") pod \"nova-cell0-conductor-db-sync-5rcr4\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.369383 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-scripts\") pod \"nova-cell0-conductor-db-sync-5rcr4\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.369558 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-5rcr4\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.369605 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-config-data\") pod \"nova-cell0-conductor-db-sync-5rcr4\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.390898 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-config-data\") pod \"nova-cell0-conductor-db-sync-5rcr4\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.394351 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-5rcr4\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.399658 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-scripts\") pod \"nova-cell0-conductor-db-sync-5rcr4\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.412588 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md6xw\" (UniqueName: \"kubernetes.io/projected/abf0902b-bd70-49df-966b-4cabf47f75df-kube-api-access-md6xw\") pod \"nova-cell0-conductor-db-sync-5rcr4\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.556154 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.727518 4777 generic.go:334] "Generic (PLEG): container finished" podID="c0eeb390-6ffc-41c8-a816-0b4454730e07" containerID="06afe8fe5773f3cf973f33c455ae72d9dda6f08c0d265ad6e15cbfc57f4376e5" exitCode=137 Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.727755 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0eeb390-6ffc-41c8-a816-0b4454730e07","Type":"ContainerDied","Data":"06afe8fe5773f3cf973f33c455ae72d9dda6f08c0d265ad6e15cbfc57f4376e5"} Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.737105 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"2aa8a135-f57b-4d6b-9444-4fbdfc542476","Type":"ContainerStarted","Data":"896237ee47f1b5107998401ff73591a93018704bd939c2fb9289492ce4fbcd36"} Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.737155 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"2aa8a135-f57b-4d6b-9444-4fbdfc542476","Type":"ContainerStarted","Data":"87cdfb2999f0d2da241c0e6943520c119fe0205538a1047748a3216a36a7ebd0"} Nov 24 17:22:23 crc kubenswrapper[4777]: I1124 17:22:23.768368 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=2.7683455 podStartE2EDuration="2.7683455s" podCreationTimestamp="2025-11-24 17:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:23.759590577 +0000 UTC m=+1301.918785626" watchObservedRunningTime="2025-11-24 17:22:23.7683455 +0000 UTC m=+1301.927540549" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.129291 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.188414 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5rcr4"] Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.294199 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-config-data\") pod \"c0eeb390-6ffc-41c8-a816-0b4454730e07\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.294529 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-combined-ca-bundle\") pod \"c0eeb390-6ffc-41c8-a816-0b4454730e07\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.294626 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-sg-core-conf-yaml\") pod \"c0eeb390-6ffc-41c8-a816-0b4454730e07\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.294675 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-scripts\") pod \"c0eeb390-6ffc-41c8-a816-0b4454730e07\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.294870 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd7sx\" (UniqueName: \"kubernetes.io/projected/c0eeb390-6ffc-41c8-a816-0b4454730e07-kube-api-access-qd7sx\") pod \"c0eeb390-6ffc-41c8-a816-0b4454730e07\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.294910 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0eeb390-6ffc-41c8-a816-0b4454730e07-run-httpd\") pod \"c0eeb390-6ffc-41c8-a816-0b4454730e07\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.294984 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0eeb390-6ffc-41c8-a816-0b4454730e07-log-httpd\") pod \"c0eeb390-6ffc-41c8-a816-0b4454730e07\" (UID: \"c0eeb390-6ffc-41c8-a816-0b4454730e07\") " Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.297273 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0eeb390-6ffc-41c8-a816-0b4454730e07-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c0eeb390-6ffc-41c8-a816-0b4454730e07" (UID: "c0eeb390-6ffc-41c8-a816-0b4454730e07"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.299032 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0eeb390-6ffc-41c8-a816-0b4454730e07-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c0eeb390-6ffc-41c8-a816-0b4454730e07" (UID: "c0eeb390-6ffc-41c8-a816-0b4454730e07"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.305097 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-scripts" (OuterVolumeSpecName: "scripts") pod "c0eeb390-6ffc-41c8-a816-0b4454730e07" (UID: "c0eeb390-6ffc-41c8-a816-0b4454730e07"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.308062 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c0eeb390-6ffc-41c8-a816-0b4454730e07" (UID: "c0eeb390-6ffc-41c8-a816-0b4454730e07"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.327358 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0eeb390-6ffc-41c8-a816-0b4454730e07-kube-api-access-qd7sx" (OuterVolumeSpecName: "kube-api-access-qd7sx") pod "c0eeb390-6ffc-41c8-a816-0b4454730e07" (UID: "c0eeb390-6ffc-41c8-a816-0b4454730e07"). InnerVolumeSpecName "kube-api-access-qd7sx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.353752 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0eeb390-6ffc-41c8-a816-0b4454730e07" (UID: "c0eeb390-6ffc-41c8-a816-0b4454730e07"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.400133 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.400159 4777 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.400169 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.400177 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd7sx\" (UniqueName: \"kubernetes.io/projected/c0eeb390-6ffc-41c8-a816-0b4454730e07-kube-api-access-qd7sx\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.400185 4777 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0eeb390-6ffc-41c8-a816-0b4454730e07-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.400194 4777 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0eeb390-6ffc-41c8-a816-0b4454730e07-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.430781 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-config-data" (OuterVolumeSpecName: "config-data") pod "c0eeb390-6ffc-41c8-a816-0b4454730e07" (UID: "c0eeb390-6ffc-41c8-a816-0b4454730e07"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.501602 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0eeb390-6ffc-41c8-a816-0b4454730e07-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.750801 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0eeb390-6ffc-41c8-a816-0b4454730e07","Type":"ContainerDied","Data":"8341686b30fa5d5d17576f224281e1d58883e70b693778743130240748380fbc"} Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.751241 4777 scope.go:117] "RemoveContainer" containerID="06afe8fe5773f3cf973f33c455ae72d9dda6f08c0d265ad6e15cbfc57f4376e5" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.751400 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.764228 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13","Type":"ContainerStarted","Data":"6a5e3a7a4beb8b3a0ae5d34b9b1a9c3d7ef822ca25c9a91d5cbd3f0389d1ba13"} Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.764284 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13","Type":"ContainerStarted","Data":"f0fcefcf1003cb5562c36cac96f0b66fcd938b5f93e977702c8cd3252927d025"} Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.771176 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5rcr4" event={"ID":"abf0902b-bd70-49df-966b-4cabf47f75df","Type":"ContainerStarted","Data":"6b868a52b46da2c8dbed42921ddef2698101785b43b2489902e77ebe419382b2"} Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.785850 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.785833484 podStartE2EDuration="3.785833484s" podCreationTimestamp="2025-11-24 17:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:24.783815015 +0000 UTC m=+1302.943010074" watchObservedRunningTime="2025-11-24 17:22:24.785833484 +0000 UTC m=+1302.945028533" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.809418 4777 scope.go:117] "RemoveContainer" containerID="28cd9ebd6d1385ee8cc3d47a866c0c2c36fbee6fda63ce837a511c472a6be062" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.856245 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.876029 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.901024 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:22:24 crc kubenswrapper[4777]: E1124 17:22:24.901467 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0eeb390-6ffc-41c8-a816-0b4454730e07" containerName="ceilometer-notification-agent" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.901485 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0eeb390-6ffc-41c8-a816-0b4454730e07" containerName="ceilometer-notification-agent" Nov 24 17:22:24 crc kubenswrapper[4777]: E1124 17:22:24.901522 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0eeb390-6ffc-41c8-a816-0b4454730e07" containerName="proxy-httpd" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.901529 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0eeb390-6ffc-41c8-a816-0b4454730e07" containerName="proxy-httpd" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.901775 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0eeb390-6ffc-41c8-a816-0b4454730e07" containerName="ceilometer-notification-agent" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.901795 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0eeb390-6ffc-41c8-a816-0b4454730e07" containerName="proxy-httpd" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.903698 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.910345 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.910517 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 17:22:24 crc kubenswrapper[4777]: I1124 17:22:24.929728 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.015428 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.015808 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-run-httpd\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.015843 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2v9w\" (UniqueName: \"kubernetes.io/projected/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-kube-api-access-q2v9w\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.015871 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-config-data\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.015980 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.016012 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-scripts\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.016156 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-log-httpd\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.122093 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.122161 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-run-httpd\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.122190 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2v9w\" (UniqueName: \"kubernetes.io/projected/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-kube-api-access-q2v9w\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.122211 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-config-data\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.122249 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.122268 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-scripts\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.122338 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-log-httpd\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.122765 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-log-httpd\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.126772 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-run-httpd\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.148662 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-scripts\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.149027 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-config-data\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.149243 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.153551 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.165785 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2v9w\" (UniqueName: \"kubernetes.io/projected/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-kube-api-access-q2v9w\") pod \"ceilometer-0\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.241156 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.310253 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0eeb390-6ffc-41c8-a816-0b4454730e07" path="/var/lib/kubelet/pods/c0eeb390-6ffc-41c8-a816-0b4454730e07/volumes" Nov 24 17:22:25 crc kubenswrapper[4777]: I1124 17:22:25.930202 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:22:25 crc kubenswrapper[4777]: W1124 17:22:25.942418 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04ff04ad_5aaf_4c00_a407_3d85ef99f69a.slice/crio-bc1888982541f9a7db33ec5a58205fa12aa87f569682d9390d91358fd3afb5e7 WatchSource:0}: Error finding container bc1888982541f9a7db33ec5a58205fa12aa87f569682d9390d91358fd3afb5e7: Status 404 returned error can't find the container with id bc1888982541f9a7db33ec5a58205fa12aa87f569682d9390d91358fd3afb5e7 Nov 24 17:22:26 crc kubenswrapper[4777]: I1124 17:22:26.864069 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04ff04ad-5aaf-4c00-a407-3d85ef99f69a","Type":"ContainerStarted","Data":"bc1888982541f9a7db33ec5a58205fa12aa87f569682d9390d91358fd3afb5e7"} Nov 24 17:22:27 crc kubenswrapper[4777]: I1124 17:22:27.164688 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 17:22:27 crc kubenswrapper[4777]: I1124 17:22:27.877646 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04ff04ad-5aaf-4c00-a407-3d85ef99f69a","Type":"ContainerStarted","Data":"cf5252ff0924b4759f7c4d2f7c540cec01d8296c964e3300b23760e1be754dc0"} Nov 24 17:22:27 crc kubenswrapper[4777]: I1124 17:22:27.877688 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04ff04ad-5aaf-4c00-a407-3d85ef99f69a","Type":"ContainerStarted","Data":"a38d9aeca402c9e55bbd68e0e7dca775400a2920f4638bf86db280f9c053a19d"} Nov 24 17:22:28 crc kubenswrapper[4777]: I1124 17:22:28.888583 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04ff04ad-5aaf-4c00-a407-3d85ef99f69a","Type":"ContainerStarted","Data":"0afae80951785f5da0b356ede0d2b012eba0361f6067812f59734b652620e73e"} Nov 24 17:22:29 crc kubenswrapper[4777]: I1124 17:22:29.728313 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:22:31 crc kubenswrapper[4777]: I1124 17:22:31.920958 4777 generic.go:334] "Generic (PLEG): container finished" podID="48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" containerID="a35e4c4bffb7897a92ffc9257c05d0e6aaab089dac5efc537ec10649c3f9d6bb" exitCode=137 Nov 24 17:22:31 crc kubenswrapper[4777]: I1124 17:22:31.921004 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd","Type":"ContainerDied","Data":"a35e4c4bffb7897a92ffc9257c05d0e6aaab089dac5efc537ec10649c3f9d6bb"} Nov 24 17:22:32 crc kubenswrapper[4777]: I1124 17:22:32.436463 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 17:22:32 crc kubenswrapper[4777]: I1124 17:22:32.560764 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.187:8776/healthcheck\": dial tcp 10.217.0.187:8776: connect: connection refused" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.581750 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.637581 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtv7m\" (UniqueName: \"kubernetes.io/projected/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-kube-api-access-qtv7m\") pod \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.637852 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-combined-ca-bundle\") pod \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.638007 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-config-data-custom\") pod \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.638122 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-etc-machine-id\") pod \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.638262 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-scripts\") pod \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.638463 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-logs\") pod \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.638585 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-config-data\") pod \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\" (UID: \"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd\") " Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.645083 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" (UID: "48a5b3ec-e2fe-46d6-af20-2b441a9a27bd"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.651297 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-logs" (OuterVolumeSpecName: "logs") pod "48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" (UID: "48a5b3ec-e2fe-46d6-af20-2b441a9a27bd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.652596 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-kube-api-access-qtv7m" (OuterVolumeSpecName: "kube-api-access-qtv7m") pod "48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" (UID: "48a5b3ec-e2fe-46d6-af20-2b441a9a27bd"). InnerVolumeSpecName "kube-api-access-qtv7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.656671 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" (UID: "48a5b3ec-e2fe-46d6-af20-2b441a9a27bd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.665783 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-scripts" (OuterVolumeSpecName: "scripts") pod "48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" (UID: "48a5b3ec-e2fe-46d6-af20-2b441a9a27bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.689115 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" (UID: "48a5b3ec-e2fe-46d6-af20-2b441a9a27bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.743346 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.743809 4777 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.743896 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtv7m\" (UniqueName: \"kubernetes.io/projected/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-kube-api-access-qtv7m\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.743955 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.744027 4777 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.744092 4777 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.743814 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-config-data" (OuterVolumeSpecName: "config-data") pod "48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" (UID: "48a5b3ec-e2fe-46d6-af20-2b441a9a27bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.851741 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.963689 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"48a5b3ec-e2fe-46d6-af20-2b441a9a27bd","Type":"ContainerDied","Data":"3e7623921478a8218fa92f576b3cc5a9882605cb60d8989e60443de476a88571"} Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.963738 4777 scope.go:117] "RemoveContainer" containerID="a35e4c4bffb7897a92ffc9257c05d0e6aaab089dac5efc537ec10649c3f9d6bb" Nov 24 17:22:35 crc kubenswrapper[4777]: I1124 17:22:35.963866 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.008189 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.017654 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.033744 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:22:36 crc kubenswrapper[4777]: E1124 17:22:36.037493 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" containerName="cinder-api-log" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.037621 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" containerName="cinder-api-log" Nov 24 17:22:36 crc kubenswrapper[4777]: E1124 17:22:36.037776 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" containerName="cinder-api" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.037850 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" containerName="cinder-api" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.038220 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" containerName="cinder-api-log" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.038338 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" containerName="cinder-api" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.039897 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.046050 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.050628 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.052220 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.052359 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.157098 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-config-data\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.157165 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20f43f58-1756-48e1-a958-b49bc82ef1b4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.157304 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-config-data-custom\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.157537 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9n8h\" (UniqueName: \"kubernetes.io/projected/20f43f58-1756-48e1-a958-b49bc82ef1b4-kube-api-access-n9n8h\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.157620 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-scripts\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.157657 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.157955 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20f43f58-1756-48e1-a958-b49bc82ef1b4-logs\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.158147 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.158267 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.259851 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9n8h\" (UniqueName: \"kubernetes.io/projected/20f43f58-1756-48e1-a958-b49bc82ef1b4-kube-api-access-n9n8h\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.259905 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-scripts\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.259931 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.259955 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20f43f58-1756-48e1-a958-b49bc82ef1b4-logs\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.259992 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.260012 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.260064 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-config-data\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.260109 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20f43f58-1756-48e1-a958-b49bc82ef1b4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.260142 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-config-data-custom\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.260460 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20f43f58-1756-48e1-a958-b49bc82ef1b4-logs\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.260921 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20f43f58-1756-48e1-a958-b49bc82ef1b4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.266692 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.267886 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-config-data\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.279397 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-scripts\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.279573 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-config-data-custom\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.282441 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.282541 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20f43f58-1756-48e1-a958-b49bc82ef1b4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.296638 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9n8h\" (UniqueName: \"kubernetes.io/projected/20f43f58-1756-48e1-a958-b49bc82ef1b4-kube-api-access-n9n8h\") pod \"cinder-api-0\" (UID: \"20f43f58-1756-48e1-a958-b49bc82ef1b4\") " pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.382077 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 17:22:36 crc kubenswrapper[4777]: I1124 17:22:36.709695 4777 scope.go:117] "RemoveContainer" containerID="5401ef1add4a6a05820876e814f93d33402154c4d4da21203a0b3c72b4b89611" Nov 24 17:22:37 crc kubenswrapper[4777]: I1124 17:22:37.275415 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48a5b3ec-e2fe-46d6-af20-2b441a9a27bd" path="/var/lib/kubelet/pods/48a5b3ec-e2fe-46d6-af20-2b441a9a27bd/volumes" Nov 24 17:22:37 crc kubenswrapper[4777]: I1124 17:22:37.753986 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 17:22:37 crc kubenswrapper[4777]: I1124 17:22:37.988352 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5rcr4" event={"ID":"abf0902b-bd70-49df-966b-4cabf47f75df","Type":"ContainerStarted","Data":"9d28c6485b14552d3905d31afaa413159ddb8d4913b720e081bc801683048732"} Nov 24 17:22:37 crc kubenswrapper[4777]: I1124 17:22:37.992762 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04ff04ad-5aaf-4c00-a407-3d85ef99f69a","Type":"ContainerStarted","Data":"204508eaff3e436ef5e79ba421e05db866c69e54be769659dd9c49a43704cf0b"} Nov 24 17:22:37 crc kubenswrapper[4777]: I1124 17:22:37.992955 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="ceilometer-central-agent" containerID="cri-o://a38d9aeca402c9e55bbd68e0e7dca775400a2920f4638bf86db280f9c053a19d" gracePeriod=30 Nov 24 17:22:37 crc kubenswrapper[4777]: I1124 17:22:37.994594 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 17:22:37 crc kubenswrapper[4777]: I1124 17:22:37.994666 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="proxy-httpd" containerID="cri-o://204508eaff3e436ef5e79ba421e05db866c69e54be769659dd9c49a43704cf0b" gracePeriod=30 Nov 24 17:22:37 crc kubenswrapper[4777]: I1124 17:22:37.994749 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="sg-core" containerID="cri-o://0afae80951785f5da0b356ede0d2b012eba0361f6067812f59734b652620e73e" gracePeriod=30 Nov 24 17:22:37 crc kubenswrapper[4777]: I1124 17:22:37.994800 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="ceilometer-notification-agent" containerID="cri-o://cf5252ff0924b4759f7c4d2f7c540cec01d8296c964e3300b23760e1be754dc0" gracePeriod=30 Nov 24 17:22:37 crc kubenswrapper[4777]: I1124 17:22:37.998621 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"20f43f58-1756-48e1-a958-b49bc82ef1b4","Type":"ContainerStarted","Data":"1c12131e1c39a21e9456f00914ef8c56de1fd03b7129c1275c2ec9b9d581de60"} Nov 24 17:22:38 crc kubenswrapper[4777]: I1124 17:22:38.008428 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-5rcr4" podStartSLOduration=1.985948871 podStartE2EDuration="15.008385652s" podCreationTimestamp="2025-11-24 17:22:23 +0000 UTC" firstStartedPulling="2025-11-24 17:22:24.205188353 +0000 UTC m=+1302.364383392" lastFinishedPulling="2025-11-24 17:22:37.227625124 +0000 UTC m=+1315.386820173" observedRunningTime="2025-11-24 17:22:38.003581413 +0000 UTC m=+1316.162776462" watchObservedRunningTime="2025-11-24 17:22:38.008385652 +0000 UTC m=+1316.167580711" Nov 24 17:22:38 crc kubenswrapper[4777]: I1124 17:22:38.039400 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.753519745 podStartE2EDuration="14.039374188s" podCreationTimestamp="2025-11-24 17:22:24 +0000 UTC" firstStartedPulling="2025-11-24 17:22:25.945277752 +0000 UTC m=+1304.104472801" lastFinishedPulling="2025-11-24 17:22:37.231132165 +0000 UTC m=+1315.390327244" observedRunningTime="2025-11-24 17:22:38.02804216 +0000 UTC m=+1316.187237239" watchObservedRunningTime="2025-11-24 17:22:38.039374188 +0000 UTC m=+1316.198569237" Nov 24 17:22:39 crc kubenswrapper[4777]: I1124 17:22:39.009837 4777 generic.go:334] "Generic (PLEG): container finished" podID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerID="0afae80951785f5da0b356ede0d2b012eba0361f6067812f59734b652620e73e" exitCode=2 Nov 24 17:22:39 crc kubenswrapper[4777]: I1124 17:22:39.010152 4777 generic.go:334] "Generic (PLEG): container finished" podID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerID="cf5252ff0924b4759f7c4d2f7c540cec01d8296c964e3300b23760e1be754dc0" exitCode=0 Nov 24 17:22:39 crc kubenswrapper[4777]: I1124 17:22:39.010164 4777 generic.go:334] "Generic (PLEG): container finished" podID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerID="a38d9aeca402c9e55bbd68e0e7dca775400a2920f4638bf86db280f9c053a19d" exitCode=0 Nov 24 17:22:39 crc kubenswrapper[4777]: I1124 17:22:39.009911 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04ff04ad-5aaf-4c00-a407-3d85ef99f69a","Type":"ContainerDied","Data":"0afae80951785f5da0b356ede0d2b012eba0361f6067812f59734b652620e73e"} Nov 24 17:22:39 crc kubenswrapper[4777]: I1124 17:22:39.010220 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04ff04ad-5aaf-4c00-a407-3d85ef99f69a","Type":"ContainerDied","Data":"cf5252ff0924b4759f7c4d2f7c540cec01d8296c964e3300b23760e1be754dc0"} Nov 24 17:22:39 crc kubenswrapper[4777]: I1124 17:22:39.010231 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04ff04ad-5aaf-4c00-a407-3d85ef99f69a","Type":"ContainerDied","Data":"a38d9aeca402c9e55bbd68e0e7dca775400a2920f4638bf86db280f9c053a19d"} Nov 24 17:22:39 crc kubenswrapper[4777]: I1124 17:22:39.015348 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"20f43f58-1756-48e1-a958-b49bc82ef1b4","Type":"ContainerStarted","Data":"5fb5d044ec2da260b1b93c63e947627b85ea6ed07cd156dbbd7e7b5e720eabc7"} Nov 24 17:22:39 crc kubenswrapper[4777]: I1124 17:22:39.496089 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:22:39 crc kubenswrapper[4777]: I1124 17:22:39.496690 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5b3890d7-b149-4be6-bff8-60f8b7de9e22" containerName="glance-log" containerID="cri-o://42977bcbcf9da3b11a6cd2bdd47f71900fec096419ad4bb2584edc54cb20e9be" gracePeriod=30 Nov 24 17:22:39 crc kubenswrapper[4777]: I1124 17:22:39.496805 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5b3890d7-b149-4be6-bff8-60f8b7de9e22" containerName="glance-httpd" containerID="cri-o://a230424fa31831eca45d0d49503ee0856c035c927587fe4e5f9f93e71fc722f0" gracePeriod=30 Nov 24 17:22:40 crc kubenswrapper[4777]: I1124 17:22:40.031508 4777 generic.go:334] "Generic (PLEG): container finished" podID="5b3890d7-b149-4be6-bff8-60f8b7de9e22" containerID="42977bcbcf9da3b11a6cd2bdd47f71900fec096419ad4bb2584edc54cb20e9be" exitCode=143 Nov 24 17:22:40 crc kubenswrapper[4777]: I1124 17:22:40.031891 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5b3890d7-b149-4be6-bff8-60f8b7de9e22","Type":"ContainerDied","Data":"42977bcbcf9da3b11a6cd2bdd47f71900fec096419ad4bb2584edc54cb20e9be"} Nov 24 17:22:40 crc kubenswrapper[4777]: I1124 17:22:40.035842 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"20f43f58-1756-48e1-a958-b49bc82ef1b4","Type":"ContainerStarted","Data":"a80f1ce665bcef3b48a4f3b21a05afcb0e71fbbabaf5feca20c53cce950533c4"} Nov 24 17:22:40 crc kubenswrapper[4777]: I1124 17:22:40.036008 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 17:22:40 crc kubenswrapper[4777]: I1124 17:22:40.059932 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.059912717 podStartE2EDuration="4.059912717s" podCreationTimestamp="2025-11-24 17:22:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:40.053248345 +0000 UTC m=+1318.212443384" watchObservedRunningTime="2025-11-24 17:22:40.059912717 +0000 UTC m=+1318.219107766" Nov 24 17:22:40 crc kubenswrapper[4777]: I1124 17:22:40.828993 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:22:40 crc kubenswrapper[4777]: I1124 17:22:40.829355 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d03c2407-e0ad-49da-91a9-c9d3e8e67692" containerName="glance-log" containerID="cri-o://fdfe37b10df787951845c17e6d8025051d9dd87fe20c768b6d1a1a5b901f0777" gracePeriod=30 Nov 24 17:22:40 crc kubenswrapper[4777]: I1124 17:22:40.829484 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d03c2407-e0ad-49da-91a9-c9d3e8e67692" containerName="glance-httpd" containerID="cri-o://26f51bd0e101b47349c5d126e029079ae8159da608e1677a07fad76ed55d5e5c" gracePeriod=30 Nov 24 17:22:41 crc kubenswrapper[4777]: I1124 17:22:41.045648 4777 generic.go:334] "Generic (PLEG): container finished" podID="d03c2407-e0ad-49da-91a9-c9d3e8e67692" containerID="fdfe37b10df787951845c17e6d8025051d9dd87fe20c768b6d1a1a5b901f0777" exitCode=143 Nov 24 17:22:41 crc kubenswrapper[4777]: I1124 17:22:41.045721 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d03c2407-e0ad-49da-91a9-c9d3e8e67692","Type":"ContainerDied","Data":"fdfe37b10df787951845c17e6d8025051d9dd87fe20c768b6d1a1a5b901f0777"} Nov 24 17:22:43 crc kubenswrapper[4777]: I1124 17:22:43.065324 4777 generic.go:334] "Generic (PLEG): container finished" podID="5b3890d7-b149-4be6-bff8-60f8b7de9e22" containerID="a230424fa31831eca45d0d49503ee0856c035c927587fe4e5f9f93e71fc722f0" exitCode=0 Nov 24 17:22:43 crc kubenswrapper[4777]: I1124 17:22:43.065389 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5b3890d7-b149-4be6-bff8-60f8b7de9e22","Type":"ContainerDied","Data":"a230424fa31831eca45d0d49503ee0856c035c927587fe4e5f9f93e71fc722f0"} Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.012015 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.090153 4777 generic.go:334] "Generic (PLEG): container finished" podID="d03c2407-e0ad-49da-91a9-c9d3e8e67692" containerID="26f51bd0e101b47349c5d126e029079ae8159da608e1677a07fad76ed55d5e5c" exitCode=0 Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.090205 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d03c2407-e0ad-49da-91a9-c9d3e8e67692","Type":"ContainerDied","Data":"26f51bd0e101b47349c5d126e029079ae8159da608e1677a07fad76ed55d5e5c"} Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.106714 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5b3890d7-b149-4be6-bff8-60f8b7de9e22","Type":"ContainerDied","Data":"72433a3f6fdb73cecb9f1ff2e4d32735952d2974d6c98d8033a7821ff9d83cc1"} Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.106761 4777 scope.go:117] "RemoveContainer" containerID="a230424fa31831eca45d0d49503ee0856c035c927587fe4e5f9f93e71fc722f0" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.106959 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.135110 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szvz2\" (UniqueName: \"kubernetes.io/projected/5b3890d7-b149-4be6-bff8-60f8b7de9e22-kube-api-access-szvz2\") pod \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.135180 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-config-data\") pod \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.135271 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5b3890d7-b149-4be6-bff8-60f8b7de9e22-httpd-run\") pod \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.135319 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-combined-ca-bundle\") pod \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.135385 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-public-tls-certs\") pod \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.135532 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") pod \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.135579 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-scripts\") pod \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.135620 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b3890d7-b149-4be6-bff8-60f8b7de9e22-logs\") pod \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.136398 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b3890d7-b149-4be6-bff8-60f8b7de9e22-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5b3890d7-b149-4be6-bff8-60f8b7de9e22" (UID: "5b3890d7-b149-4be6-bff8-60f8b7de9e22"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.136912 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b3890d7-b149-4be6-bff8-60f8b7de9e22-logs" (OuterVolumeSpecName: "logs") pod "5b3890d7-b149-4be6-bff8-60f8b7de9e22" (UID: "5b3890d7-b149-4be6-bff8-60f8b7de9e22"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.137585 4777 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b3890d7-b149-4be6-bff8-60f8b7de9e22-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.137611 4777 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5b3890d7-b149-4be6-bff8-60f8b7de9e22-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.142255 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b3890d7-b149-4be6-bff8-60f8b7de9e22-kube-api-access-szvz2" (OuterVolumeSpecName: "kube-api-access-szvz2") pod "5b3890d7-b149-4be6-bff8-60f8b7de9e22" (UID: "5b3890d7-b149-4be6-bff8-60f8b7de9e22"). InnerVolumeSpecName "kube-api-access-szvz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.160319 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-scripts" (OuterVolumeSpecName: "scripts") pod "5b3890d7-b149-4be6-bff8-60f8b7de9e22" (UID: "5b3890d7-b149-4be6-bff8-60f8b7de9e22"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: E1124 17:22:44.176267 4777 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8 podName:5b3890d7-b149-4be6-bff8-60f8b7de9e22 nodeName:}" failed. No retries permitted until 2025-11-24 17:22:44.676247993 +0000 UTC m=+1322.835443042 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "glance" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8") pod "5b3890d7-b149-4be6-bff8-60f8b7de9e22" (UID: "5b3890d7-b149-4be6-bff8-60f8b7de9e22") : kubernetes.io/csi: Unmounter.TearDownAt failed: rpc error: code = Unknown desc = check target path: could not get consistent content of /proc/mounts after 3 attempts Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.182182 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b3890d7-b149-4be6-bff8-60f8b7de9e22" (UID: "5b3890d7-b149-4be6-bff8-60f8b7de9e22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.213132 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5b3890d7-b149-4be6-bff8-60f8b7de9e22" (UID: "5b3890d7-b149-4be6-bff8-60f8b7de9e22"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.238216 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-config-data" (OuterVolumeSpecName: "config-data") pod "5b3890d7-b149-4be6-bff8-60f8b7de9e22" (UID: "5b3890d7-b149-4be6-bff8-60f8b7de9e22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.238443 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-config-data\") pod \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.239274 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szvz2\" (UniqueName: \"kubernetes.io/projected/5b3890d7-b149-4be6-bff8-60f8b7de9e22-kube-api-access-szvz2\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.239295 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.239304 4777 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.239311 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:44 crc kubenswrapper[4777]: W1124 17:22:44.239388 4777 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/5b3890d7-b149-4be6-bff8-60f8b7de9e22/volumes/kubernetes.io~secret/config-data Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.239397 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-config-data" (OuterVolumeSpecName: "config-data") pod "5b3890d7-b149-4be6-bff8-60f8b7de9e22" (UID: "5b3890d7-b149-4be6-bff8-60f8b7de9e22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.264855 4777 scope.go:117] "RemoveContainer" containerID="42977bcbcf9da3b11a6cd2bdd47f71900fec096419ad4bb2584edc54cb20e9be" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.270581 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.270807 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.270837 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.271373 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a6faf411f1a27fb01d7bc7ffb8b05ff8c95006e6030cf4cec387d5e3b99354f9"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.271419 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://a6faf411f1a27fb01d7bc7ffb8b05ff8c95006e6030cf4cec387d5e3b99354f9" gracePeriod=600 Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.341706 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b3890d7-b149-4be6-bff8-60f8b7de9e22-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.545384 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.646262 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d03c2407-e0ad-49da-91a9-c9d3e8e67692-logs\") pod \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.646405 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d03c2407-e0ad-49da-91a9-c9d3e8e67692-httpd-run\") pod \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.646461 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m95pf\" (UniqueName: \"kubernetes.io/projected/d03c2407-e0ad-49da-91a9-c9d3e8e67692-kube-api-access-m95pf\") pod \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.646536 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") pod \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.646571 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-config-data\") pod \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.646592 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-scripts\") pod \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.646610 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-internal-tls-certs\") pod \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.646662 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-combined-ca-bundle\") pod \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\" (UID: \"d03c2407-e0ad-49da-91a9-c9d3e8e67692\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.649454 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d03c2407-e0ad-49da-91a9-c9d3e8e67692-logs" (OuterVolumeSpecName: "logs") pod "d03c2407-e0ad-49da-91a9-c9d3e8e67692" (UID: "d03c2407-e0ad-49da-91a9-c9d3e8e67692"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.649704 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d03c2407-e0ad-49da-91a9-c9d3e8e67692-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d03c2407-e0ad-49da-91a9-c9d3e8e67692" (UID: "d03c2407-e0ad-49da-91a9-c9d3e8e67692"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.653002 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-scripts" (OuterVolumeSpecName: "scripts") pod "d03c2407-e0ad-49da-91a9-c9d3e8e67692" (UID: "d03c2407-e0ad-49da-91a9-c9d3e8e67692"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.658057 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d03c2407-e0ad-49da-91a9-c9d3e8e67692-kube-api-access-m95pf" (OuterVolumeSpecName: "kube-api-access-m95pf") pod "d03c2407-e0ad-49da-91a9-c9d3e8e67692" (UID: "d03c2407-e0ad-49da-91a9-c9d3e8e67692"). InnerVolumeSpecName "kube-api-access-m95pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.682925 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4" (OuterVolumeSpecName: "glance") pod "d03c2407-e0ad-49da-91a9-c9d3e8e67692" (UID: "d03c2407-e0ad-49da-91a9-c9d3e8e67692"). InnerVolumeSpecName "pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.694138 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d03c2407-e0ad-49da-91a9-c9d3e8e67692" (UID: "d03c2407-e0ad-49da-91a9-c9d3e8e67692"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.733697 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-config-data" (OuterVolumeSpecName: "config-data") pod "d03c2407-e0ad-49da-91a9-c9d3e8e67692" (UID: "d03c2407-e0ad-49da-91a9-c9d3e8e67692"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.735703 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d03c2407-e0ad-49da-91a9-c9d3e8e67692" (UID: "d03c2407-e0ad-49da-91a9-c9d3e8e67692"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.748799 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") pod \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\" (UID: \"5b3890d7-b149-4be6-bff8-60f8b7de9e22\") " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.749401 4777 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d03c2407-e0ad-49da-91a9-c9d3e8e67692-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.749428 4777 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d03c2407-e0ad-49da-91a9-c9d3e8e67692-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.749439 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m95pf\" (UniqueName: \"kubernetes.io/projected/d03c2407-e0ad-49da-91a9-c9d3e8e67692-kube-api-access-m95pf\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.749469 4777 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") on node \"crc\" " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.749482 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.749493 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.749504 4777 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.749516 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03c2407-e0ad-49da-91a9-c9d3e8e67692-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.763997 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8" (OuterVolumeSpecName: "glance") pod "5b3890d7-b149-4be6-bff8-60f8b7de9e22" (UID: "5b3890d7-b149-4be6-bff8-60f8b7de9e22"). InnerVolumeSpecName "pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.772537 4777 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.772678 4777 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4") on node "crc" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.851790 4777 reconciler_common.go:293] "Volume detached for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.851845 4777 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") on node \"crc\" " Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.867370 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.878136 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.890835 4777 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.890995 4777 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8") on node "crc" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.891589 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:22:44 crc kubenswrapper[4777]: E1124 17:22:44.892024 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b3890d7-b149-4be6-bff8-60f8b7de9e22" containerName="glance-log" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.892043 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b3890d7-b149-4be6-bff8-60f8b7de9e22" containerName="glance-log" Nov 24 17:22:44 crc kubenswrapper[4777]: E1124 17:22:44.892061 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b3890d7-b149-4be6-bff8-60f8b7de9e22" containerName="glance-httpd" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.892067 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b3890d7-b149-4be6-bff8-60f8b7de9e22" containerName="glance-httpd" Nov 24 17:22:44 crc kubenswrapper[4777]: E1124 17:22:44.892098 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d03c2407-e0ad-49da-91a9-c9d3e8e67692" containerName="glance-httpd" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.892106 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="d03c2407-e0ad-49da-91a9-c9d3e8e67692" containerName="glance-httpd" Nov 24 17:22:44 crc kubenswrapper[4777]: E1124 17:22:44.892139 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d03c2407-e0ad-49da-91a9-c9d3e8e67692" containerName="glance-log" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.892145 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="d03c2407-e0ad-49da-91a9-c9d3e8e67692" containerName="glance-log" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.892353 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="d03c2407-e0ad-49da-91a9-c9d3e8e67692" containerName="glance-log" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.892370 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="d03c2407-e0ad-49da-91a9-c9d3e8e67692" containerName="glance-httpd" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.892394 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b3890d7-b149-4be6-bff8-60f8b7de9e22" containerName="glance-log" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.892406 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b3890d7-b149-4be6-bff8-60f8b7de9e22" containerName="glance-httpd" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.893860 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.896585 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.898444 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.901340 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.959366 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-scripts\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.960450 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.960559 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.960650 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv5cj\" (UniqueName: \"kubernetes.io/projected/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-kube-api-access-hv5cj\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.960773 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-logs\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.960854 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-config-data\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.960944 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.961065 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.968239 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:22:44 crc kubenswrapper[4777]: I1124 17:22:44.968316 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/23594f9d6cc83f7e0c53346caabd46c775bdd206833b4ed7a43e80540ba27612/globalmount\"" pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.016221 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0fea9df0-8b5c-4519-9a1a-42f55f94efc8\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.063149 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-scripts\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.063196 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.063234 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.063274 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv5cj\" (UniqueName: \"kubernetes.io/projected/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-kube-api-access-hv5cj\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.063302 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-logs\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.063333 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-config-data\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.063370 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.064175 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-logs\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.068423 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.070716 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.070876 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-config-data\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.096323 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-scripts\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.096666 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.102103 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv5cj\" (UniqueName: \"kubernetes.io/projected/44214c39-ae9a-4e14-a7ec-fc1e95051d7f-kube-api-access-hv5cj\") pod \"glance-default-external-api-0\" (UID: \"44214c39-ae9a-4e14-a7ec-fc1e95051d7f\") " pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.124079 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d03c2407-e0ad-49da-91a9-c9d3e8e67692","Type":"ContainerDied","Data":"e2e6dceb91b23e0437e91788b4a923a59cadd2eccf302e3c4b729355fc4ee9b8"} Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.124195 4777 scope.go:117] "RemoveContainer" containerID="26f51bd0e101b47349c5d126e029079ae8159da608e1677a07fad76ed55d5e5c" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.124345 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.136748 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="a6faf411f1a27fb01d7bc7ffb8b05ff8c95006e6030cf4cec387d5e3b99354f9" exitCode=0 Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.136787 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"a6faf411f1a27fb01d7bc7ffb8b05ff8c95006e6030cf4cec387d5e3b99354f9"} Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.195516 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.206685 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.224106 4777 scope.go:117] "RemoveContainer" containerID="fdfe37b10df787951845c17e6d8025051d9dd87fe20c768b6d1a1a5b901f0777" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.227929 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.242485 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.242595 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.246697 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.248656 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.278095 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.296344 4777 scope.go:117] "RemoveContainer" containerID="3cf02e0241e501ffd35b75c9b5e93243f3fad1bfd50b46e463488c0fec7f7c53" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.316439 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b3890d7-b149-4be6-bff8-60f8b7de9e22" path="/var/lib/kubelet/pods/5b3890d7-b149-4be6-bff8-60f8b7de9e22/volumes" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.317247 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d03c2407-e0ad-49da-91a9-c9d3e8e67692" path="/var/lib/kubelet/pods/d03c2407-e0ad-49da-91a9-c9d3e8e67692/volumes" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.371220 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.371719 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.371807 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.371875 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7658\" (UniqueName: \"kubernetes.io/projected/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-kube-api-access-t7658\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.372008 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-logs\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.372085 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.372153 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.372245 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.473745 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.473855 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.473905 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.473934 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7658\" (UniqueName: \"kubernetes.io/projected/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-kube-api-access-t7658\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.473987 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-logs\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.474010 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.474032 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.474056 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.478264 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-logs\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.478376 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.485009 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.485082 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.485223 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.485396 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.486733 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.486763 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0d3411b018636661d0f6478efc7fafc27f5aef76a9aca82320677e283d3b0a36/globalmount\"" pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.503209 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7658\" (UniqueName: \"kubernetes.io/projected/b698d8b2-ca7b-43e7-8da9-aef7d3ba1538-kube-api-access-t7658\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.564007 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bfd8180-b3c5-434d-bfdc-54f8c75dd4e4\") pod \"glance-default-internal-api-0\" (UID: \"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538\") " pod="openstack/glance-default-internal-api-0" Nov 24 17:22:45 crc kubenswrapper[4777]: I1124 17:22:45.591555 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 17:22:46 crc kubenswrapper[4777]: I1124 17:22:46.055147 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 17:22:46 crc kubenswrapper[4777]: I1124 17:22:46.146829 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"44214c39-ae9a-4e14-a7ec-fc1e95051d7f","Type":"ContainerStarted","Data":"a6ba876d3ba8f1ddd3637932e62bf4aebae41be5d4740e826f1607adcc08b26b"} Nov 24 17:22:46 crc kubenswrapper[4777]: I1124 17:22:46.152479 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4"} Nov 24 17:22:46 crc kubenswrapper[4777]: I1124 17:22:46.285535 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 17:22:47 crc kubenswrapper[4777]: I1124 17:22:47.167791 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538","Type":"ContainerStarted","Data":"234b8618c86218c3bf4b61f6d9ca10f2cd5ed345f6f79e9fbb62e14df693e5cb"} Nov 24 17:22:47 crc kubenswrapper[4777]: I1124 17:22:47.168478 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538","Type":"ContainerStarted","Data":"2fa437e7176bd990b7ab31bf1523c760c622cc4abf278491e0e4608a23f4193c"} Nov 24 17:22:47 crc kubenswrapper[4777]: I1124 17:22:47.170117 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"44214c39-ae9a-4e14-a7ec-fc1e95051d7f","Type":"ContainerStarted","Data":"3b4bdc03714aa664f86aa3d117e651d7c2a633adaa42df79e69e59dd40302230"} Nov 24 17:22:48 crc kubenswrapper[4777]: I1124 17:22:48.183060 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b698d8b2-ca7b-43e7-8da9-aef7d3ba1538","Type":"ContainerStarted","Data":"40383ca2b9094b670cb6422d04f1af0e5e8a76a0dbadbbe300a82d3480b6ee11"} Nov 24 17:22:48 crc kubenswrapper[4777]: I1124 17:22:48.185317 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"44214c39-ae9a-4e14-a7ec-fc1e95051d7f","Type":"ContainerStarted","Data":"35b631b0b942b0c2e8d3bc1f2ea4c3e49be8663759a45eb30ae5ecf407b67487"} Nov 24 17:22:48 crc kubenswrapper[4777]: I1124 17:22:48.210615 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.210586367 podStartE2EDuration="3.210586367s" podCreationTimestamp="2025-11-24 17:22:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:48.204512342 +0000 UTC m=+1326.363707391" watchObservedRunningTime="2025-11-24 17:22:48.210586367 +0000 UTC m=+1326.369781456" Nov 24 17:22:48 crc kubenswrapper[4777]: I1124 17:22:48.237308 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.237283539 podStartE2EDuration="4.237283539s" podCreationTimestamp="2025-11-24 17:22:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:48.229045641 +0000 UTC m=+1326.388240690" watchObservedRunningTime="2025-11-24 17:22:48.237283539 +0000 UTC m=+1326.396478588" Nov 24 17:22:48 crc kubenswrapper[4777]: I1124 17:22:48.728524 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 17:22:55 crc kubenswrapper[4777]: I1124 17:22:55.268458 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 17:22:55 crc kubenswrapper[4777]: I1124 17:22:55.278242 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 17:22:55 crc kubenswrapper[4777]: I1124 17:22:55.278293 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 17:22:55 crc kubenswrapper[4777]: I1124 17:22:55.283879 4777 generic.go:334] "Generic (PLEG): container finished" podID="abf0902b-bd70-49df-966b-4cabf47f75df" containerID="9d28c6485b14552d3905d31afaa413159ddb8d4913b720e081bc801683048732" exitCode=0 Nov 24 17:22:55 crc kubenswrapper[4777]: I1124 17:22:55.283927 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5rcr4" event={"ID":"abf0902b-bd70-49df-966b-4cabf47f75df","Type":"ContainerDied","Data":"9d28c6485b14552d3905d31afaa413159ddb8d4913b720e081bc801683048732"} Nov 24 17:22:55 crc kubenswrapper[4777]: I1124 17:22:55.320597 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 17:22:55 crc kubenswrapper[4777]: I1124 17:22:55.324261 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 17:22:55 crc kubenswrapper[4777]: I1124 17:22:55.391548 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-api-0" Nov 24 17:22:55 crc kubenswrapper[4777]: I1124 17:22:55.591875 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 17:22:55 crc kubenswrapper[4777]: I1124 17:22:55.592243 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 17:22:55 crc kubenswrapper[4777]: I1124 17:22:55.627182 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 17:22:55 crc kubenswrapper[4777]: I1124 17:22:55.636779 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 17:22:56 crc kubenswrapper[4777]: I1124 17:22:56.294188 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 17:22:56 crc kubenswrapper[4777]: I1124 17:22:56.294224 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 17:22:56 crc kubenswrapper[4777]: I1124 17:22:56.294234 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 17:22:56 crc kubenswrapper[4777]: I1124 17:22:56.294242 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 17:22:56 crc kubenswrapper[4777]: I1124 17:22:56.882518 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.023549 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md6xw\" (UniqueName: \"kubernetes.io/projected/abf0902b-bd70-49df-966b-4cabf47f75df-kube-api-access-md6xw\") pod \"abf0902b-bd70-49df-966b-4cabf47f75df\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.023706 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-combined-ca-bundle\") pod \"abf0902b-bd70-49df-966b-4cabf47f75df\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.023799 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-scripts\") pod \"abf0902b-bd70-49df-966b-4cabf47f75df\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.023930 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-config-data\") pod \"abf0902b-bd70-49df-966b-4cabf47f75df\" (UID: \"abf0902b-bd70-49df-966b-4cabf47f75df\") " Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.053125 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-scripts" (OuterVolumeSpecName: "scripts") pod "abf0902b-bd70-49df-966b-4cabf47f75df" (UID: "abf0902b-bd70-49df-966b-4cabf47f75df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.055694 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abf0902b-bd70-49df-966b-4cabf47f75df" (UID: "abf0902b-bd70-49df-966b-4cabf47f75df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.118872 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abf0902b-bd70-49df-966b-4cabf47f75df-kube-api-access-md6xw" (OuterVolumeSpecName: "kube-api-access-md6xw") pod "abf0902b-bd70-49df-966b-4cabf47f75df" (UID: "abf0902b-bd70-49df-966b-4cabf47f75df"). InnerVolumeSpecName "kube-api-access-md6xw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.125115 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-config-data" (OuterVolumeSpecName: "config-data") pod "abf0902b-bd70-49df-966b-4cabf47f75df" (UID: "abf0902b-bd70-49df-966b-4cabf47f75df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.126528 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md6xw\" (UniqueName: \"kubernetes.io/projected/abf0902b-bd70-49df-966b-4cabf47f75df-kube-api-access-md6xw\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.126639 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.126706 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.126773 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abf0902b-bd70-49df-966b-4cabf47f75df-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.306890 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5rcr4" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.307345 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5rcr4" event={"ID":"abf0902b-bd70-49df-966b-4cabf47f75df","Type":"ContainerDied","Data":"6b868a52b46da2c8dbed42921ddef2698101785b43b2489902e77ebe419382b2"} Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.307371 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b868a52b46da2c8dbed42921ddef2698101785b43b2489902e77ebe419382b2" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.431182 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 17:22:57 crc kubenswrapper[4777]: E1124 17:22:57.431595 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abf0902b-bd70-49df-966b-4cabf47f75df" containerName="nova-cell0-conductor-db-sync" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.431608 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="abf0902b-bd70-49df-966b-4cabf47f75df" containerName="nova-cell0-conductor-db-sync" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.431779 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="abf0902b-bd70-49df-966b-4cabf47f75df" containerName="nova-cell0-conductor-db-sync" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.432500 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.435261 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-npnv4" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.435984 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.446375 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.541650 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aee763a5-b286-4fdd-9573-d06a0e915f9e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"aee763a5-b286-4fdd-9573-d06a0e915f9e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.541715 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cb5r\" (UniqueName: \"kubernetes.io/projected/aee763a5-b286-4fdd-9573-d06a0e915f9e-kube-api-access-8cb5r\") pod \"nova-cell0-conductor-0\" (UID: \"aee763a5-b286-4fdd-9573-d06a0e915f9e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.541762 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aee763a5-b286-4fdd-9573-d06a0e915f9e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"aee763a5-b286-4fdd-9573-d06a0e915f9e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.644042 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aee763a5-b286-4fdd-9573-d06a0e915f9e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"aee763a5-b286-4fdd-9573-d06a0e915f9e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.644322 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aee763a5-b286-4fdd-9573-d06a0e915f9e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"aee763a5-b286-4fdd-9573-d06a0e915f9e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.644366 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cb5r\" (UniqueName: \"kubernetes.io/projected/aee763a5-b286-4fdd-9573-d06a0e915f9e-kube-api-access-8cb5r\") pod \"nova-cell0-conductor-0\" (UID: \"aee763a5-b286-4fdd-9573-d06a0e915f9e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.858698 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aee763a5-b286-4fdd-9573-d06a0e915f9e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"aee763a5-b286-4fdd-9573-d06a0e915f9e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.858846 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aee763a5-b286-4fdd-9573-d06a0e915f9e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"aee763a5-b286-4fdd-9573-d06a0e915f9e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:22:57 crc kubenswrapper[4777]: I1124 17:22:57.858858 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cb5r\" (UniqueName: \"kubernetes.io/projected/aee763a5-b286-4fdd-9573-d06a0e915f9e-kube-api-access-8cb5r\") pod \"nova-cell0-conductor-0\" (UID: \"aee763a5-b286-4fdd-9573-d06a0e915f9e\") " pod="openstack/nova-cell0-conductor-0" Nov 24 17:22:58 crc kubenswrapper[4777]: I1124 17:22:58.052373 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 17:22:58 crc kubenswrapper[4777]: I1124 17:22:58.710330 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 17:22:58 crc kubenswrapper[4777]: I1124 17:22:58.759112 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 17:22:58 crc kubenswrapper[4777]: I1124 17:22:58.759218 4777 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 17:22:58 crc kubenswrapper[4777]: I1124 17:22:58.765579 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 17:22:59 crc kubenswrapper[4777]: I1124 17:22:59.368110 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"aee763a5-b286-4fdd-9573-d06a0e915f9e","Type":"ContainerStarted","Data":"642d8b11fc66d5cb7093d08c5ffff84534558e32bde27316590699b344c26b69"} Nov 24 17:22:59 crc kubenswrapper[4777]: I1124 17:22:59.368529 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"aee763a5-b286-4fdd-9573-d06a0e915f9e","Type":"ContainerStarted","Data":"6d90db9729fc675df74251bc3e4cd52309baa3e1f81d0eeec3fcac4d737227c1"} Nov 24 17:22:59 crc kubenswrapper[4777]: I1124 17:22:59.368556 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 17:22:59 crc kubenswrapper[4777]: I1124 17:22:59.385854 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.385835762 podStartE2EDuration="2.385835762s" podCreationTimestamp="2025-11-24 17:22:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:22:59.383085203 +0000 UTC m=+1337.542280262" watchObservedRunningTime="2025-11-24 17:22:59.385835762 +0000 UTC m=+1337.545030811" Nov 24 17:22:59 crc kubenswrapper[4777]: I1124 17:22:59.510289 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 17:22:59 crc kubenswrapper[4777]: I1124 17:22:59.510392 4777 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 17:22:59 crc kubenswrapper[4777]: I1124 17:22:59.518219 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 17:23:08 crc kubenswrapper[4777]: I1124 17:23:08.106572 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 17:23:08 crc kubenswrapper[4777]: E1124 17:23:08.332677 4777 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04ff04ad_5aaf_4c00_a407_3d85ef99f69a.slice/crio-conmon-204508eaff3e436ef5e79ba421e05db866c69e54be769659dd9c49a43704cf0b.scope\": RecentStats: unable to find data in memory cache]" Nov 24 17:23:08 crc kubenswrapper[4777]: I1124 17:23:08.465417 4777 generic.go:334] "Generic (PLEG): container finished" podID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerID="204508eaff3e436ef5e79ba421e05db866c69e54be769659dd9c49a43704cf0b" exitCode=137 Nov 24 17:23:08 crc kubenswrapper[4777]: I1124 17:23:08.465483 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04ff04ad-5aaf-4c00-a407-3d85ef99f69a","Type":"ContainerDied","Data":"204508eaff3e436ef5e79ba421e05db866c69e54be769659dd9c49a43704cf0b"} Nov 24 17:23:08 crc kubenswrapper[4777]: I1124 17:23:08.906390 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-vpbfr"] Nov 24 17:23:08 crc kubenswrapper[4777]: I1124 17:23:08.908892 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:08 crc kubenswrapper[4777]: I1124 17:23:08.914334 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 24 17:23:08 crc kubenswrapper[4777]: I1124 17:23:08.914346 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 24 17:23:08 crc kubenswrapper[4777]: I1124 17:23:08.927132 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vpbfr"] Nov 24 17:23:08 crc kubenswrapper[4777]: I1124 17:23:08.994982 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vpbfr\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:08 crc kubenswrapper[4777]: I1124 17:23:08.995031 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-scripts\") pod \"nova-cell0-cell-mapping-vpbfr\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:08 crc kubenswrapper[4777]: I1124 17:23:08.995064 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-config-data\") pod \"nova-cell0-cell-mapping-vpbfr\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:08 crc kubenswrapper[4777]: I1124 17:23:08.995098 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66d8g\" (UniqueName: \"kubernetes.io/projected/4366459e-a3b8-4216-9540-547eeaf67972-kube-api-access-66d8g\") pod \"nova-cell0-cell-mapping-vpbfr\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.096500 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vpbfr\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.096552 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-scripts\") pod \"nova-cell0-cell-mapping-vpbfr\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.096590 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-config-data\") pod \"nova-cell0-cell-mapping-vpbfr\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.096636 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66d8g\" (UniqueName: \"kubernetes.io/projected/4366459e-a3b8-4216-9540-547eeaf67972-kube-api-access-66d8g\") pod \"nova-cell0-cell-mapping-vpbfr\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.108002 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-config-data\") pod \"nova-cell0-cell-mapping-vpbfr\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.110596 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-scripts\") pod \"nova-cell0-cell-mapping-vpbfr\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.121826 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.123914 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.150110 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.153310 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vpbfr\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.155411 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.182132 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66d8g\" (UniqueName: \"kubernetes.io/projected/4366459e-a3b8-4216-9540-547eeaf67972-kube-api-access-66d8g\") pod \"nova-cell0-cell-mapping-vpbfr\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.201445 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5lr2\" (UniqueName: \"kubernetes.io/projected/a1fc5297-b877-4fcb-9391-becc710d2a24-kube-api-access-w5lr2\") pod \"nova-metadata-0\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " pod="openstack/nova-metadata-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.201565 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1fc5297-b877-4fcb-9391-becc710d2a24-config-data\") pod \"nova-metadata-0\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " pod="openstack/nova-metadata-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.201609 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1fc5297-b877-4fcb-9391-becc710d2a24-logs\") pod \"nova-metadata-0\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " pod="openstack/nova-metadata-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.201648 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1fc5297-b877-4fcb-9391-becc710d2a24-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " pod="openstack/nova-metadata-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.214486 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.217821 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.228313 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.235389 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.254230 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.305149 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5lr2\" (UniqueName: \"kubernetes.io/projected/a1fc5297-b877-4fcb-9391-becc710d2a24-kube-api-access-w5lr2\") pod \"nova-metadata-0\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " pod="openstack/nova-metadata-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.305263 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1fc5297-b877-4fcb-9391-becc710d2a24-config-data\") pod \"nova-metadata-0\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " pod="openstack/nova-metadata-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.305311 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1fc5297-b877-4fcb-9391-becc710d2a24-logs\") pod \"nova-metadata-0\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " pod="openstack/nova-metadata-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.305349 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1fc5297-b877-4fcb-9391-becc710d2a24-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " pod="openstack/nova-metadata-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.306377 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.307505 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.321939 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1fc5297-b877-4fcb-9391-becc710d2a24-logs\") pod \"nova-metadata-0\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " pod="openstack/nova-metadata-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.328342 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.345314 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1fc5297-b877-4fcb-9391-becc710d2a24-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " pod="openstack/nova-metadata-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.346587 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1fc5297-b877-4fcb-9391-becc710d2a24-config-data\") pod \"nova-metadata-0\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " pod="openstack/nova-metadata-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.355473 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.358768 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.361562 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5lr2\" (UniqueName: \"kubernetes.io/projected/a1fc5297-b877-4fcb-9391-becc710d2a24-kube-api-access-w5lr2\") pod \"nova-metadata-0\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " pod="openstack/nova-metadata-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.392179 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.406596 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:23:09 crc kubenswrapper[4777]: E1124 17:23:09.407083 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="ceilometer-central-agent" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.407100 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="ceilometer-central-agent" Nov 24 17:23:09 crc kubenswrapper[4777]: E1124 17:23:09.407115 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="sg-core" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.407121 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="sg-core" Nov 24 17:23:09 crc kubenswrapper[4777]: E1124 17:23:09.407132 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="ceilometer-notification-agent" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.407140 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="ceilometer-notification-agent" Nov 24 17:23:09 crc kubenswrapper[4777]: E1124 17:23:09.407148 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="proxy-httpd" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.407155 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="proxy-httpd" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.407326 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="sg-core" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.409501 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7bceb11-3897-46ea-acd5-afeee23f55a1-logs\") pod \"nova-api-0\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " pod="openstack/nova-api-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.415513 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk8f8\" (UniqueName: \"kubernetes.io/projected/67ecd2f1-6485-4d94-b66c-0bc0465ea695-kube-api-access-jk8f8\") pod \"nova-cell1-novncproxy-0\" (UID: \"67ecd2f1-6485-4d94-b66c-0bc0465ea695\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.415639 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ecd2f1-6485-4d94-b66c-0bc0465ea695-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"67ecd2f1-6485-4d94-b66c-0bc0465ea695\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.415932 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bceb11-3897-46ea-acd5-afeee23f55a1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " pod="openstack/nova-api-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.416254 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ecd2f1-6485-4d94-b66c-0bc0465ea695-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"67ecd2f1-6485-4d94-b66c-0bc0465ea695\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.416479 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7wbm\" (UniqueName: \"kubernetes.io/projected/a7bceb11-3897-46ea-acd5-afeee23f55a1-kube-api-access-b7wbm\") pod \"nova-api-0\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " pod="openstack/nova-api-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.416637 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bceb11-3897-46ea-acd5-afeee23f55a1-config-data\") pod \"nova-api-0\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " pod="openstack/nova-api-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.407344 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="proxy-httpd" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.422094 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="ceilometer-notification-agent" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.422137 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" containerName="ceilometer-central-agent" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.423360 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.445870 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.503677 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-884c8b8f5-v85nr"] Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.505472 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.518449 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-run-httpd\") pod \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.518569 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-sg-core-conf-yaml\") pod \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.518669 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-config-data\") pod \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.518763 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-scripts\") pod \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.518826 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-combined-ca-bundle\") pod \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.518920 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-log-httpd\") pod \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.519005 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2v9w\" (UniqueName: \"kubernetes.io/projected/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-kube-api-access-q2v9w\") pod \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\" (UID: \"04ff04ad-5aaf-4c00-a407-3d85ef99f69a\") " Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.519323 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ecd2f1-6485-4d94-b66c-0bc0465ea695-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"67ecd2f1-6485-4d94-b66c-0bc0465ea695\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.519411 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67dde531-8482-4ab8-9da0-b07999fde7fb-config-data\") pod \"nova-scheduler-0\" (UID: \"67dde531-8482-4ab8-9da0-b07999fde7fb\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.519430 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7wbm\" (UniqueName: \"kubernetes.io/projected/a7bceb11-3897-46ea-acd5-afeee23f55a1-kube-api-access-b7wbm\") pod \"nova-api-0\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " pod="openstack/nova-api-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.519486 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bceb11-3897-46ea-acd5-afeee23f55a1-config-data\") pod \"nova-api-0\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " pod="openstack/nova-api-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.519513 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67dde531-8482-4ab8-9da0-b07999fde7fb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"67dde531-8482-4ab8-9da0-b07999fde7fb\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.519633 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z8ln\" (UniqueName: \"kubernetes.io/projected/67dde531-8482-4ab8-9da0-b07999fde7fb-kube-api-access-9z8ln\") pod \"nova-scheduler-0\" (UID: \"67dde531-8482-4ab8-9da0-b07999fde7fb\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.519934 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7bceb11-3897-46ea-acd5-afeee23f55a1-logs\") pod \"nova-api-0\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " pod="openstack/nova-api-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.520122 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk8f8\" (UniqueName: \"kubernetes.io/projected/67ecd2f1-6485-4d94-b66c-0bc0465ea695-kube-api-access-jk8f8\") pod \"nova-cell1-novncproxy-0\" (UID: \"67ecd2f1-6485-4d94-b66c-0bc0465ea695\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.520142 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ecd2f1-6485-4d94-b66c-0bc0465ea695-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"67ecd2f1-6485-4d94-b66c-0bc0465ea695\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.520226 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bceb11-3897-46ea-acd5-afeee23f55a1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " pod="openstack/nova-api-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.522092 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "04ff04ad-5aaf-4c00-a407-3d85ef99f69a" (UID: "04ff04ad-5aaf-4c00-a407-3d85ef99f69a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.522656 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "04ff04ad-5aaf-4c00-a407-3d85ef99f69a" (UID: "04ff04ad-5aaf-4c00-a407-3d85ef99f69a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.526404 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.527452 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7bceb11-3897-46ea-acd5-afeee23f55a1-logs\") pod \"nova-api-0\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " pod="openstack/nova-api-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.534033 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-kube-api-access-q2v9w" (OuterVolumeSpecName: "kube-api-access-q2v9w") pod "04ff04ad-5aaf-4c00-a407-3d85ef99f69a" (UID: "04ff04ad-5aaf-4c00-a407-3d85ef99f69a"). InnerVolumeSpecName "kube-api-access-q2v9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.534050 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04ff04ad-5aaf-4c00-a407-3d85ef99f69a","Type":"ContainerDied","Data":"bc1888982541f9a7db33ec5a58205fa12aa87f569682d9390d91358fd3afb5e7"} Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.534094 4777 scope.go:117] "RemoveContainer" containerID="204508eaff3e436ef5e79ba421e05db866c69e54be769659dd9c49a43704cf0b" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.534220 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.536001 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-scripts" (OuterVolumeSpecName: "scripts") pod "04ff04ad-5aaf-4c00-a407-3d85ef99f69a" (UID: "04ff04ad-5aaf-4c00-a407-3d85ef99f69a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.537016 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ecd2f1-6485-4d94-b66c-0bc0465ea695-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"67ecd2f1-6485-4d94-b66c-0bc0465ea695\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.545845 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ecd2f1-6485-4d94-b66c-0bc0465ea695-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"67ecd2f1-6485-4d94-b66c-0bc0465ea695\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.545905 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bceb11-3897-46ea-acd5-afeee23f55a1-config-data\") pod \"nova-api-0\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " pod="openstack/nova-api-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.549460 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bceb11-3897-46ea-acd5-afeee23f55a1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " pod="openstack/nova-api-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.549602 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7wbm\" (UniqueName: \"kubernetes.io/projected/a7bceb11-3897-46ea-acd5-afeee23f55a1-kube-api-access-b7wbm\") pod \"nova-api-0\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " pod="openstack/nova-api-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.549756 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-884c8b8f5-v85nr"] Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.552791 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk8f8\" (UniqueName: \"kubernetes.io/projected/67ecd2f1-6485-4d94-b66c-0bc0465ea695-kube-api-access-jk8f8\") pod \"nova-cell1-novncproxy-0\" (UID: \"67ecd2f1-6485-4d94-b66c-0bc0465ea695\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.579607 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "04ff04ad-5aaf-4c00-a407-3d85ef99f69a" (UID: "04ff04ad-5aaf-4c00-a407-3d85ef99f69a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.596439 4777 scope.go:117] "RemoveContainer" containerID="0afae80951785f5da0b356ede0d2b012eba0361f6067812f59734b652620e73e" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.622233 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-dns-svc\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.622291 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67dde531-8482-4ab8-9da0-b07999fde7fb-config-data\") pod \"nova-scheduler-0\" (UID: \"67dde531-8482-4ab8-9da0-b07999fde7fb\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.622346 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67dde531-8482-4ab8-9da0-b07999fde7fb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"67dde531-8482-4ab8-9da0-b07999fde7fb\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.622387 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-ovsdbserver-sb\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.622414 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-dns-swift-storage-0\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.622436 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z8ln\" (UniqueName: \"kubernetes.io/projected/67dde531-8482-4ab8-9da0-b07999fde7fb-kube-api-access-9z8ln\") pod \"nova-scheduler-0\" (UID: \"67dde531-8482-4ab8-9da0-b07999fde7fb\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.622458 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcp7t\" (UniqueName: \"kubernetes.io/projected/8a758074-c834-4b50-8f0e-8959700e9cab-kube-api-access-dcp7t\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.622487 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-ovsdbserver-nb\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.622558 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-config\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.622680 4777 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.622691 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2v9w\" (UniqueName: \"kubernetes.io/projected/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-kube-api-access-q2v9w\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.622709 4777 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.622717 4777 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.622725 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.628208 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67dde531-8482-4ab8-9da0-b07999fde7fb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"67dde531-8482-4ab8-9da0-b07999fde7fb\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.636266 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67dde531-8482-4ab8-9da0-b07999fde7fb-config-data\") pod \"nova-scheduler-0\" (UID: \"67dde531-8482-4ab8-9da0-b07999fde7fb\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.662147 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z8ln\" (UniqueName: \"kubernetes.io/projected/67dde531-8482-4ab8-9da0-b07999fde7fb-kube-api-access-9z8ln\") pod \"nova-scheduler-0\" (UID: \"67dde531-8482-4ab8-9da0-b07999fde7fb\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.675489 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04ff04ad-5aaf-4c00-a407-3d85ef99f69a" (UID: "04ff04ad-5aaf-4c00-a407-3d85ef99f69a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.706633 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-config-data" (OuterVolumeSpecName: "config-data") pod "04ff04ad-5aaf-4c00-a407-3d85ef99f69a" (UID: "04ff04ad-5aaf-4c00-a407-3d85ef99f69a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.722708 4777 scope.go:117] "RemoveContainer" containerID="cf5252ff0924b4759f7c4d2f7c540cec01d8296c964e3300b23760e1be754dc0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.724165 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-dns-svc\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.724279 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-ovsdbserver-sb\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.724323 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-dns-swift-storage-0\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.724366 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcp7t\" (UniqueName: \"kubernetes.io/projected/8a758074-c834-4b50-8f0e-8959700e9cab-kube-api-access-dcp7t\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.724397 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-ovsdbserver-nb\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.724472 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-config\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.724552 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.724571 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04ff04ad-5aaf-4c00-a407-3d85ef99f69a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.725502 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-config\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.726459 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-ovsdbserver-nb\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.726676 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-dns-svc\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.727188 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-ovsdbserver-sb\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.727439 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-dns-swift-storage-0\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.748677 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcp7t\" (UniqueName: \"kubernetes.io/projected/8a758074-c834-4b50-8f0e-8959700e9cab-kube-api-access-dcp7t\") pod \"dnsmasq-dns-884c8b8f5-v85nr\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.762436 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.775606 4777 scope.go:117] "RemoveContainer" containerID="a38d9aeca402c9e55bbd68e0e7dca775400a2920f4638bf86db280f9c053a19d" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.794890 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.864380 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.882886 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.911535 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.918076 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.935044 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.937677 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.940322 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.940371 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.955028 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:23:09 crc kubenswrapper[4777]: I1124 17:23:09.975375 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vpbfr"] Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.033450 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-config-data\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.033597 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.033688 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.033720 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-scripts\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.033782 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6882541d-9663-41db-9d0a-ec16ead4f421-log-httpd\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.033858 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vcb4\" (UniqueName: \"kubernetes.io/projected/6882541d-9663-41db-9d0a-ec16ead4f421-kube-api-access-4vcb4\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.033951 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6882541d-9663-41db-9d0a-ec16ead4f421-run-httpd\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.053699 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.135955 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vcb4\" (UniqueName: \"kubernetes.io/projected/6882541d-9663-41db-9d0a-ec16ead4f421-kube-api-access-4vcb4\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.136052 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6882541d-9663-41db-9d0a-ec16ead4f421-run-httpd\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.136110 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-config-data\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.136140 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.136178 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.136198 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-scripts\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.136234 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6882541d-9663-41db-9d0a-ec16ead4f421-log-httpd\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.137007 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6882541d-9663-41db-9d0a-ec16ead4f421-log-httpd\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.138481 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6882541d-9663-41db-9d0a-ec16ead4f421-run-httpd\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.144015 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-config-data\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.144351 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.149860 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-scripts\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.156288 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vcb4\" (UniqueName: \"kubernetes.io/projected/6882541d-9663-41db-9d0a-ec16ead4f421-kube-api-access-4vcb4\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.158454 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.264382 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.394367 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:23:10 crc kubenswrapper[4777]: W1124 17:23:10.582090 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67ecd2f1_6485_4d94_b66c_0bc0465ea695.slice/crio-207c191a791bd18fd51972fd1fc5ef64c31a1d8e3c2d6732710c44002b8498db WatchSource:0}: Error finding container 207c191a791bd18fd51972fd1fc5ef64c31a1d8e3c2d6732710c44002b8498db: Status 404 returned error can't find the container with id 207c191a791bd18fd51972fd1fc5ef64c31a1d8e3c2d6732710c44002b8498db Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.587481 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.591669 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7bceb11-3897-46ea-acd5-afeee23f55a1","Type":"ContainerStarted","Data":"3958b07eb74156bdf7e9b77b216b3e307e839a042bd1b4f657a3e363b5c0d031"} Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.621171 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vpbfr" event={"ID":"4366459e-a3b8-4216-9540-547eeaf67972","Type":"ContainerStarted","Data":"8c44adf30eb15f68a0635029c88c02d72b8615d819ae6e6ee0d2957bd4d47abd"} Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.621406 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vpbfr" event={"ID":"4366459e-a3b8-4216-9540-547eeaf67972","Type":"ContainerStarted","Data":"a50928e445f02d54bbc6da37bf92bd68bdb6251f8a1608256aeff3c310f7a3ae"} Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.634760 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1fc5297-b877-4fcb-9391-becc710d2a24","Type":"ContainerStarted","Data":"c62538b895d095e34200db1095815440a34cebb31b9055e8da239bb63590a19b"} Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.653952 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-vpbfr" podStartSLOduration=2.653935662 podStartE2EDuration="2.653935662s" podCreationTimestamp="2025-11-24 17:23:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:23:10.648236787 +0000 UTC m=+1348.807431836" watchObservedRunningTime="2025-11-24 17:23:10.653935662 +0000 UTC m=+1348.813130711" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.807442 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.824221 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-884c8b8f5-v85nr"] Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.866998 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xfksm"] Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.868425 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.871491 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.871585 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.888086 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xfksm"] Nov 24 17:23:10 crc kubenswrapper[4777]: I1124 17:23:10.999927 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-config-data\") pod \"nova-cell1-conductor-db-sync-xfksm\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.000110 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9dwg\" (UniqueName: \"kubernetes.io/projected/7d534624-b898-4268-a432-7f7a575fb15b-kube-api-access-p9dwg\") pod \"nova-cell1-conductor-db-sync-xfksm\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.000168 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-scripts\") pod \"nova-cell1-conductor-db-sync-xfksm\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.000422 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xfksm\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:11 crc kubenswrapper[4777]: W1124 17:23:11.101858 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6882541d_9663_41db_9d0a_ec16ead4f421.slice/crio-591832c702a388c65a27f77490095502a9e5762742f2f7fb5d4c24effcc112da WatchSource:0}: Error finding container 591832c702a388c65a27f77490095502a9e5762742f2f7fb5d4c24effcc112da: Status 404 returned error can't find the container with id 591832c702a388c65a27f77490095502a9e5762742f2f7fb5d4c24effcc112da Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.102338 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xfksm\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.102394 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-config-data\") pod \"nova-cell1-conductor-db-sync-xfksm\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.102480 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9dwg\" (UniqueName: \"kubernetes.io/projected/7d534624-b898-4268-a432-7f7a575fb15b-kube-api-access-p9dwg\") pod \"nova-cell1-conductor-db-sync-xfksm\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.102531 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-scripts\") pod \"nova-cell1-conductor-db-sync-xfksm\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.102557 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.106739 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xfksm\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.106785 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-config-data\") pod \"nova-cell1-conductor-db-sync-xfksm\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.106831 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-scripts\") pod \"nova-cell1-conductor-db-sync-xfksm\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.126114 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9dwg\" (UniqueName: \"kubernetes.io/projected/7d534624-b898-4268-a432-7f7a575fb15b-kube-api-access-p9dwg\") pod \"nova-cell1-conductor-db-sync-xfksm\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.192890 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.294336 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04ff04ad-5aaf-4c00-a407-3d85ef99f69a" path="/var/lib/kubelet/pods/04ff04ad-5aaf-4c00-a407-3d85ef99f69a/volumes" Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.646290 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6882541d-9663-41db-9d0a-ec16ead4f421","Type":"ContainerStarted","Data":"591832c702a388c65a27f77490095502a9e5762742f2f7fb5d4c24effcc112da"} Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.650447 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"67dde531-8482-4ab8-9da0-b07999fde7fb","Type":"ContainerStarted","Data":"80a6ccaa65db195566ee0fb259cba0c784db433d6d5e85a86707b9f93f3fee3d"} Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.654364 4777 generic.go:334] "Generic (PLEG): container finished" podID="8a758074-c834-4b50-8f0e-8959700e9cab" containerID="6747216930bd90bd130aea78411607661ebcfdfe42532a16dfb7485da87dd047" exitCode=0 Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.654433 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" event={"ID":"8a758074-c834-4b50-8f0e-8959700e9cab","Type":"ContainerDied","Data":"6747216930bd90bd130aea78411607661ebcfdfe42532a16dfb7485da87dd047"} Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.654458 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" event={"ID":"8a758074-c834-4b50-8f0e-8959700e9cab","Type":"ContainerStarted","Data":"d936cddb2620ace12eee0a90c826a590079b5e1eebabc84ceb41cbb9985ed9ef"} Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.661993 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"67ecd2f1-6485-4d94-b66c-0bc0465ea695","Type":"ContainerStarted","Data":"207c191a791bd18fd51972fd1fc5ef64c31a1d8e3c2d6732710c44002b8498db"} Nov 24 17:23:11 crc kubenswrapper[4777]: I1124 17:23:11.787466 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xfksm"] Nov 24 17:23:11 crc kubenswrapper[4777]: W1124 17:23:11.798501 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d534624_b898_4268_a432_7f7a575fb15b.slice/crio-702d6e67650ae256ab18f6f334c074ea12f91c91becfecdc7b4b561d6d3b3bf9 WatchSource:0}: Error finding container 702d6e67650ae256ab18f6f334c074ea12f91c91becfecdc7b4b561d6d3b3bf9: Status 404 returned error can't find the container with id 702d6e67650ae256ab18f6f334c074ea12f91c91becfecdc7b4b561d6d3b3bf9 Nov 24 17:23:12 crc kubenswrapper[4777]: I1124 17:23:12.673404 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xfksm" event={"ID":"7d534624-b898-4268-a432-7f7a575fb15b","Type":"ContainerStarted","Data":"702d6e67650ae256ab18f6f334c074ea12f91c91becfecdc7b4b561d6d3b3bf9"} Nov 24 17:23:13 crc kubenswrapper[4777]: I1124 17:23:13.003392 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:23:13 crc kubenswrapper[4777]: I1124 17:23:13.029686 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:23:13 crc kubenswrapper[4777]: I1124 17:23:13.685819 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6882541d-9663-41db-9d0a-ec16ead4f421","Type":"ContainerStarted","Data":"6e14cb85635d3e244139a9e247c2bce46931d36cabc3d9bd13250d204420dda6"} Nov 24 17:23:13 crc kubenswrapper[4777]: I1124 17:23:13.689421 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" event={"ID":"8a758074-c834-4b50-8f0e-8959700e9cab","Type":"ContainerStarted","Data":"287001c4ae3dcc885df3c7ccad15a8be02ef267473fdfc38ee67728123ab4169"} Nov 24 17:23:13 crc kubenswrapper[4777]: I1124 17:23:13.690596 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:13 crc kubenswrapper[4777]: I1124 17:23:13.700570 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xfksm" event={"ID":"7d534624-b898-4268-a432-7f7a575fb15b","Type":"ContainerStarted","Data":"2c0a4fc5e29a49f0e4e2d0e4b5ffd7824db5cf109cd192786408a0ac53f33e7a"} Nov 24 17:23:13 crc kubenswrapper[4777]: I1124 17:23:13.714120 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" podStartSLOduration=4.714103766 podStartE2EDuration="4.714103766s" podCreationTimestamp="2025-11-24 17:23:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:23:13.709696908 +0000 UTC m=+1351.868891967" watchObservedRunningTime="2025-11-24 17:23:13.714103766 +0000 UTC m=+1351.873298815" Nov 24 17:23:13 crc kubenswrapper[4777]: I1124 17:23:13.739250 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-xfksm" podStartSLOduration=3.739228992 podStartE2EDuration="3.739228992s" podCreationTimestamp="2025-11-24 17:23:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:23:13.727100961 +0000 UTC m=+1351.886296010" watchObservedRunningTime="2025-11-24 17:23:13.739228992 +0000 UTC m=+1351.898424041" Nov 24 17:23:14 crc kubenswrapper[4777]: I1124 17:23:14.409500 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="d03c2407-e0ad-49da-91a9-c9d3e8e67692" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.173:9292/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 17:23:14 crc kubenswrapper[4777]: I1124 17:23:14.409910 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="d03c2407-e0ad-49da-91a9-c9d3e8e67692" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.173:9292/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.781829 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6882541d-9663-41db-9d0a-ec16ead4f421","Type":"ContainerStarted","Data":"29d2470b7cadd938824606893257bf5cda27a59d52c06dea02db705e7e6d80ba"} Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.782489 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6882541d-9663-41db-9d0a-ec16ead4f421","Type":"ContainerStarted","Data":"c6aed0abf3b15443476bb3d78d14298f34006b46dd8b8b47d5549d7272fb324c"} Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.784742 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"67dde531-8482-4ab8-9da0-b07999fde7fb","Type":"ContainerStarted","Data":"222cac5b4bc9425b5bf197e8939456f1fa7bf2cd2ad09ca3b679b9bd52f61842"} Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.786664 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"67ecd2f1-6485-4d94-b66c-0bc0465ea695","Type":"ContainerStarted","Data":"f58ae0ff9d205c07f2d8d7a420739188b36a3777636f74796eec5753453b8abd"} Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.786922 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="67ecd2f1-6485-4d94-b66c-0bc0465ea695" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://f58ae0ff9d205c07f2d8d7a420739188b36a3777636f74796eec5753453b8abd" gracePeriod=30 Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.788956 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1fc5297-b877-4fcb-9391-becc710d2a24","Type":"ContainerStarted","Data":"c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea"} Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.789157 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1fc5297-b877-4fcb-9391-becc710d2a24","Type":"ContainerStarted","Data":"cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53"} Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.789081 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a1fc5297-b877-4fcb-9391-becc710d2a24" containerName="nova-metadata-metadata" containerID="cri-o://c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea" gracePeriod=30 Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.789040 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a1fc5297-b877-4fcb-9391-becc710d2a24" containerName="nova-metadata-log" containerID="cri-o://cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53" gracePeriod=30 Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.792343 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7bceb11-3897-46ea-acd5-afeee23f55a1","Type":"ContainerStarted","Data":"959e2278072dd12b5d8f2c584e32b52da47ef209b072babfd7e5344a973e4cfe"} Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.792575 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7bceb11-3897-46ea-acd5-afeee23f55a1","Type":"ContainerStarted","Data":"3b269dd6b8f7d4e44841d81bf8334b719036d567161d55cad4818734cc18a246"} Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.795354 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.818878 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.052887447 podStartE2EDuration="10.818855492s" podCreationTimestamp="2025-11-24 17:23:09 +0000 UTC" firstStartedPulling="2025-11-24 17:23:10.852443362 +0000 UTC m=+1349.011638411" lastFinishedPulling="2025-11-24 17:23:18.618411367 +0000 UTC m=+1356.777606456" observedRunningTime="2025-11-24 17:23:19.803683433 +0000 UTC m=+1357.962878492" watchObservedRunningTime="2025-11-24 17:23:19.818855492 +0000 UTC m=+1357.978050551" Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.843492 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.718194107 podStartE2EDuration="10.843466513s" podCreationTimestamp="2025-11-24 17:23:09 +0000 UTC" firstStartedPulling="2025-11-24 17:23:10.451182749 +0000 UTC m=+1348.610377798" lastFinishedPulling="2025-11-24 17:23:18.576455155 +0000 UTC m=+1356.735650204" observedRunningTime="2025-11-24 17:23:19.828405738 +0000 UTC m=+1357.987600797" watchObservedRunningTime="2025-11-24 17:23:19.843466513 +0000 UTC m=+1358.002661572" Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.859214 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.351447142 podStartE2EDuration="10.859193148s" podCreationTimestamp="2025-11-24 17:23:09 +0000 UTC" firstStartedPulling="2025-11-24 17:23:10.072283142 +0000 UTC m=+1348.231478181" lastFinishedPulling="2025-11-24 17:23:18.580029098 +0000 UTC m=+1356.739224187" observedRunningTime="2025-11-24 17:23:19.851553867 +0000 UTC m=+1358.010748936" watchObservedRunningTime="2025-11-24 17:23:19.859193148 +0000 UTC m=+1358.018388207" Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.865179 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.865227 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.876110 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.848867587 podStartE2EDuration="10.876087537s" podCreationTimestamp="2025-11-24 17:23:09 +0000 UTC" firstStartedPulling="2025-11-24 17:23:10.591111045 +0000 UTC m=+1348.750306094" lastFinishedPulling="2025-11-24 17:23:18.618330995 +0000 UTC m=+1356.777526044" observedRunningTime="2025-11-24 17:23:19.869393793 +0000 UTC m=+1358.028588852" watchObservedRunningTime="2025-11-24 17:23:19.876087537 +0000 UTC m=+1358.035282596" Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.912179 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.915230 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.988260 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58bd69657f-bbqzr"] Nov 24 17:23:19 crc kubenswrapper[4777]: I1124 17:23:19.988500 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" podUID="2933da9d-6e63-4382-9e7c-6fc6437a7d1c" containerName="dnsmasq-dns" containerID="cri-o://51e4bec409efb6b7f41777f798933b79c2faf6cf256bb0117bef7083b7b0380e" gracePeriod=10 Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.741208 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.748394 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.826447 4777 generic.go:334] "Generic (PLEG): container finished" podID="2933da9d-6e63-4382-9e7c-6fc6437a7d1c" containerID="51e4bec409efb6b7f41777f798933b79c2faf6cf256bb0117bef7083b7b0380e" exitCode=0 Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.826631 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" event={"ID":"2933da9d-6e63-4382-9e7c-6fc6437a7d1c","Type":"ContainerDied","Data":"51e4bec409efb6b7f41777f798933b79c2faf6cf256bb0117bef7083b7b0380e"} Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.826704 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.826857 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" event={"ID":"2933da9d-6e63-4382-9e7c-6fc6437a7d1c","Type":"ContainerDied","Data":"5878fec6149fb6de39210826d6e51240fe56734dd8ed3f2d44c70adc241a93cb"} Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.826918 4777 scope.go:117] "RemoveContainer" containerID="51e4bec409efb6b7f41777f798933b79c2faf6cf256bb0117bef7083b7b0380e" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.836209 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1fc5297-b877-4fcb-9391-becc710d2a24-logs\") pod \"a1fc5297-b877-4fcb-9391-becc710d2a24\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.836246 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-dns-svc\") pod \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.836333 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-ovsdbserver-sb\") pod \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.836403 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-dns-swift-storage-0\") pod \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.836429 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-ovsdbserver-nb\") pod \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.836453 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1fc5297-b877-4fcb-9391-becc710d2a24-combined-ca-bundle\") pod \"a1fc5297-b877-4fcb-9391-becc710d2a24\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.836487 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-config\") pod \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.836507 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1fc5297-b877-4fcb-9391-becc710d2a24-config-data\") pod \"a1fc5297-b877-4fcb-9391-becc710d2a24\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.836561 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5lr2\" (UniqueName: \"kubernetes.io/projected/a1fc5297-b877-4fcb-9391-becc710d2a24-kube-api-access-w5lr2\") pod \"a1fc5297-b877-4fcb-9391-becc710d2a24\" (UID: \"a1fc5297-b877-4fcb-9391-becc710d2a24\") " Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.836620 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b889l\" (UniqueName: \"kubernetes.io/projected/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-kube-api-access-b889l\") pod \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\" (UID: \"2933da9d-6e63-4382-9e7c-6fc6437a7d1c\") " Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.840336 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1fc5297-b877-4fcb-9391-becc710d2a24-logs" (OuterVolumeSpecName: "logs") pod "a1fc5297-b877-4fcb-9391-becc710d2a24" (UID: "a1fc5297-b877-4fcb-9391-becc710d2a24"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.852301 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1fc5297-b877-4fcb-9391-becc710d2a24-kube-api-access-w5lr2" (OuterVolumeSpecName: "kube-api-access-w5lr2") pod "a1fc5297-b877-4fcb-9391-becc710d2a24" (UID: "a1fc5297-b877-4fcb-9391-becc710d2a24"). InnerVolumeSpecName "kube-api-access-w5lr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.856514 4777 generic.go:334] "Generic (PLEG): container finished" podID="a1fc5297-b877-4fcb-9391-becc710d2a24" containerID="c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea" exitCode=0 Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.856541 4777 generic.go:334] "Generic (PLEG): container finished" podID="a1fc5297-b877-4fcb-9391-becc710d2a24" containerID="cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53" exitCode=143 Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.856692 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1fc5297-b877-4fcb-9391-becc710d2a24","Type":"ContainerDied","Data":"c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea"} Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.856747 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1fc5297-b877-4fcb-9391-becc710d2a24","Type":"ContainerDied","Data":"cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53"} Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.856757 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1fc5297-b877-4fcb-9391-becc710d2a24","Type":"ContainerDied","Data":"c62538b895d095e34200db1095815440a34cebb31b9055e8da239bb63590a19b"} Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.856827 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.872035 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-kube-api-access-b889l" (OuterVolumeSpecName: "kube-api-access-b889l") pod "2933da9d-6e63-4382-9e7c-6fc6437a7d1c" (UID: "2933da9d-6e63-4382-9e7c-6fc6437a7d1c"). InnerVolumeSpecName "kube-api-access-b889l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.900110 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1fc5297-b877-4fcb-9391-becc710d2a24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1fc5297-b877-4fcb-9391-becc710d2a24" (UID: "a1fc5297-b877-4fcb-9391-becc710d2a24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.945880 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1fc5297-b877-4fcb-9391-becc710d2a24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.945921 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5lr2\" (UniqueName: \"kubernetes.io/projected/a1fc5297-b877-4fcb-9391-becc710d2a24-kube-api-access-w5lr2\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.945936 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b889l\" (UniqueName: \"kubernetes.io/projected/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-kube-api-access-b889l\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.945948 4777 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1fc5297-b877-4fcb-9391-becc710d2a24-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.948653 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2933da9d-6e63-4382-9e7c-6fc6437a7d1c" (UID: "2933da9d-6e63-4382-9e7c-6fc6437a7d1c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.965617 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.965767 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2933da9d-6e63-4382-9e7c-6fc6437a7d1c" (UID: "2933da9d-6e63-4382-9e7c-6fc6437a7d1c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.980026 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1fc5297-b877-4fcb-9391-becc710d2a24-config-data" (OuterVolumeSpecName: "config-data") pod "a1fc5297-b877-4fcb-9391-becc710d2a24" (UID: "a1fc5297-b877-4fcb-9391-becc710d2a24"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:20 crc kubenswrapper[4777]: I1124 17:23:20.992745 4777 scope.go:117] "RemoveContainer" containerID="ca52531873febeefc73dc4b0c52b73511302b230f603103fa06605719d3bc668" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.012856 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2933da9d-6e63-4382-9e7c-6fc6437a7d1c" (UID: "2933da9d-6e63-4382-9e7c-6fc6437a7d1c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.028438 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2933da9d-6e63-4382-9e7c-6fc6437a7d1c" (UID: "2933da9d-6e63-4382-9e7c-6fc6437a7d1c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.030654 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-config" (OuterVolumeSpecName: "config") pod "2933da9d-6e63-4382-9e7c-6fc6437a7d1c" (UID: "2933da9d-6e63-4382-9e7c-6fc6437a7d1c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.044305 4777 scope.go:117] "RemoveContainer" containerID="51e4bec409efb6b7f41777f798933b79c2faf6cf256bb0117bef7083b7b0380e" Nov 24 17:23:21 crc kubenswrapper[4777]: E1124 17:23:21.044740 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51e4bec409efb6b7f41777f798933b79c2faf6cf256bb0117bef7083b7b0380e\": container with ID starting with 51e4bec409efb6b7f41777f798933b79c2faf6cf256bb0117bef7083b7b0380e not found: ID does not exist" containerID="51e4bec409efb6b7f41777f798933b79c2faf6cf256bb0117bef7083b7b0380e" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.044780 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51e4bec409efb6b7f41777f798933b79c2faf6cf256bb0117bef7083b7b0380e"} err="failed to get container status \"51e4bec409efb6b7f41777f798933b79c2faf6cf256bb0117bef7083b7b0380e\": rpc error: code = NotFound desc = could not find container \"51e4bec409efb6b7f41777f798933b79c2faf6cf256bb0117bef7083b7b0380e\": container with ID starting with 51e4bec409efb6b7f41777f798933b79c2faf6cf256bb0117bef7083b7b0380e not found: ID does not exist" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.044807 4777 scope.go:117] "RemoveContainer" containerID="ca52531873febeefc73dc4b0c52b73511302b230f603103fa06605719d3bc668" Nov 24 17:23:21 crc kubenswrapper[4777]: E1124 17:23:21.045129 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca52531873febeefc73dc4b0c52b73511302b230f603103fa06605719d3bc668\": container with ID starting with ca52531873febeefc73dc4b0c52b73511302b230f603103fa06605719d3bc668 not found: ID does not exist" containerID="ca52531873febeefc73dc4b0c52b73511302b230f603103fa06605719d3bc668" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.045150 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca52531873febeefc73dc4b0c52b73511302b230f603103fa06605719d3bc668"} err="failed to get container status \"ca52531873febeefc73dc4b0c52b73511302b230f603103fa06605719d3bc668\": rpc error: code = NotFound desc = could not find container \"ca52531873febeefc73dc4b0c52b73511302b230f603103fa06605719d3bc668\": container with ID starting with ca52531873febeefc73dc4b0c52b73511302b230f603103fa06605719d3bc668 not found: ID does not exist" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.045164 4777 scope.go:117] "RemoveContainer" containerID="c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.047762 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.047783 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.047792 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1fc5297-b877-4fcb-9391-becc710d2a24-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.047801 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.047808 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.047816 4777 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2933da9d-6e63-4382-9e7c-6fc6437a7d1c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.079340 4777 scope.go:117] "RemoveContainer" containerID="cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.126135 4777 scope.go:117] "RemoveContainer" containerID="c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea" Nov 24 17:23:21 crc kubenswrapper[4777]: E1124 17:23:21.127237 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea\": container with ID starting with c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea not found: ID does not exist" containerID="c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.127278 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea"} err="failed to get container status \"c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea\": rpc error: code = NotFound desc = could not find container \"c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea\": container with ID starting with c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea not found: ID does not exist" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.127302 4777 scope.go:117] "RemoveContainer" containerID="cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53" Nov 24 17:23:21 crc kubenswrapper[4777]: E1124 17:23:21.127625 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53\": container with ID starting with cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53 not found: ID does not exist" containerID="cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.127646 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53"} err="failed to get container status \"cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53\": rpc error: code = NotFound desc = could not find container \"cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53\": container with ID starting with cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53 not found: ID does not exist" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.127661 4777 scope.go:117] "RemoveContainer" containerID="c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.127957 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea"} err="failed to get container status \"c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea\": rpc error: code = NotFound desc = could not find container \"c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea\": container with ID starting with c05be9e7f772bdbbc1fe6eef9eba74cc2c60419479a95eca65ae551e24404eea not found: ID does not exist" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.128044 4777 scope.go:117] "RemoveContainer" containerID="cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.128327 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53"} err="failed to get container status \"cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53\": rpc error: code = NotFound desc = could not find container \"cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53\": container with ID starting with cc52162495e93acd3df9b8c5b88e4bdd09e749611db4e5cb7c0969ac87497e53 not found: ID does not exist" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.194349 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58bd69657f-bbqzr"] Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.209079 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58bd69657f-bbqzr"] Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.231778 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.261617 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2933da9d-6e63-4382-9e7c-6fc6437a7d1c" path="/var/lib/kubelet/pods/2933da9d-6e63-4382-9e7c-6fc6437a7d1c/volumes" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.264610 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.266458 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:23:21 crc kubenswrapper[4777]: E1124 17:23:21.266784 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1fc5297-b877-4fcb-9391-becc710d2a24" containerName="nova-metadata-log" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.266802 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1fc5297-b877-4fcb-9391-becc710d2a24" containerName="nova-metadata-log" Nov 24 17:23:21 crc kubenswrapper[4777]: E1124 17:23:21.266815 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2933da9d-6e63-4382-9e7c-6fc6437a7d1c" containerName="init" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.266822 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="2933da9d-6e63-4382-9e7c-6fc6437a7d1c" containerName="init" Nov 24 17:23:21 crc kubenswrapper[4777]: E1124 17:23:21.266845 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2933da9d-6e63-4382-9e7c-6fc6437a7d1c" containerName="dnsmasq-dns" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.266851 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="2933da9d-6e63-4382-9e7c-6fc6437a7d1c" containerName="dnsmasq-dns" Nov 24 17:23:21 crc kubenswrapper[4777]: E1124 17:23:21.266877 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1fc5297-b877-4fcb-9391-becc710d2a24" containerName="nova-metadata-metadata" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.266884 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1fc5297-b877-4fcb-9391-becc710d2a24" containerName="nova-metadata-metadata" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.267115 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1fc5297-b877-4fcb-9391-becc710d2a24" containerName="nova-metadata-log" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.267138 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="2933da9d-6e63-4382-9e7c-6fc6437a7d1c" containerName="dnsmasq-dns" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.267151 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1fc5297-b877-4fcb-9391-becc710d2a24" containerName="nova-metadata-metadata" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.268443 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.270368 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.271520 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.283952 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.366388 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.366451 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.366580 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85940131-d100-4c1c-876e-3bea01dcc20b-logs\") pod \"nova-metadata-0\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.366670 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kqvq\" (UniqueName: \"kubernetes.io/projected/85940131-d100-4c1c-876e-3bea01dcc20b-kube-api-access-8kqvq\") pod \"nova-metadata-0\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.366738 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-config-data\") pod \"nova-metadata-0\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.469148 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kqvq\" (UniqueName: \"kubernetes.io/projected/85940131-d100-4c1c-876e-3bea01dcc20b-kube-api-access-8kqvq\") pod \"nova-metadata-0\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.469304 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-config-data\") pod \"nova-metadata-0\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.469570 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.469629 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.469690 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85940131-d100-4c1c-876e-3bea01dcc20b-logs\") pod \"nova-metadata-0\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.471702 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85940131-d100-4c1c-876e-3bea01dcc20b-logs\") pod \"nova-metadata-0\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.476235 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-config-data\") pod \"nova-metadata-0\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.476297 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.480577 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.496599 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kqvq\" (UniqueName: \"kubernetes.io/projected/85940131-d100-4c1c-876e-3bea01dcc20b-kube-api-access-8kqvq\") pod \"nova-metadata-0\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.588622 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.874491 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6882541d-9663-41db-9d0a-ec16ead4f421","Type":"ContainerStarted","Data":"6f8ee2273c0cc1a79dea4f80589dfbbdb693bba486bfd47d18b0c39f0810f381"} Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.877202 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 17:23:21 crc kubenswrapper[4777]: I1124 17:23:21.904116 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.219016111 podStartE2EDuration="12.904093982s" podCreationTimestamp="2025-11-24 17:23:09 +0000 UTC" firstStartedPulling="2025-11-24 17:23:11.108734344 +0000 UTC m=+1349.267929393" lastFinishedPulling="2025-11-24 17:23:20.793812215 +0000 UTC m=+1358.953007264" observedRunningTime="2025-11-24 17:23:21.896614345 +0000 UTC m=+1360.055809394" watchObservedRunningTime="2025-11-24 17:23:21.904093982 +0000 UTC m=+1360.063289031" Nov 24 17:23:22 crc kubenswrapper[4777]: I1124 17:23:22.465994 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:23:22 crc kubenswrapper[4777]: I1124 17:23:22.892436 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"85940131-d100-4c1c-876e-3bea01dcc20b","Type":"ContainerStarted","Data":"82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555"} Nov 24 17:23:22 crc kubenswrapper[4777]: I1124 17:23:22.892490 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"85940131-d100-4c1c-876e-3bea01dcc20b","Type":"ContainerStarted","Data":"b5fbe4365e942f99258cdf99dadc77d311064f5835444ff19db8db02c97f9a38"} Nov 24 17:23:22 crc kubenswrapper[4777]: I1124 17:23:22.894101 4777 generic.go:334] "Generic (PLEG): container finished" podID="4366459e-a3b8-4216-9540-547eeaf67972" containerID="8c44adf30eb15f68a0635029c88c02d72b8615d819ae6e6ee0d2957bd4d47abd" exitCode=0 Nov 24 17:23:22 crc kubenswrapper[4777]: I1124 17:23:22.895637 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vpbfr" event={"ID":"4366459e-a3b8-4216-9540-547eeaf67972","Type":"ContainerDied","Data":"8c44adf30eb15f68a0635029c88c02d72b8615d819ae6e6ee0d2957bd4d47abd"} Nov 24 17:23:23 crc kubenswrapper[4777]: I1124 17:23:23.256980 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1fc5297-b877-4fcb-9391-becc710d2a24" path="/var/lib/kubelet/pods/a1fc5297-b877-4fcb-9391-becc710d2a24/volumes" Nov 24 17:23:23 crc kubenswrapper[4777]: I1124 17:23:23.907237 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"85940131-d100-4c1c-876e-3bea01dcc20b","Type":"ContainerStarted","Data":"d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e"} Nov 24 17:23:23 crc kubenswrapper[4777]: I1124 17:23:23.951854 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.951831458 podStartE2EDuration="2.951831458s" podCreationTimestamp="2025-11-24 17:23:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:23:23.928570035 +0000 UTC m=+1362.087765124" watchObservedRunningTime="2025-11-24 17:23:23.951831458 +0000 UTC m=+1362.111026517" Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.426237 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.561454 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-combined-ca-bundle\") pod \"4366459e-a3b8-4216-9540-547eeaf67972\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.562672 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-scripts\") pod \"4366459e-a3b8-4216-9540-547eeaf67972\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.562794 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66d8g\" (UniqueName: \"kubernetes.io/projected/4366459e-a3b8-4216-9540-547eeaf67972-kube-api-access-66d8g\") pod \"4366459e-a3b8-4216-9540-547eeaf67972\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.562835 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-config-data\") pod \"4366459e-a3b8-4216-9540-547eeaf67972\" (UID: \"4366459e-a3b8-4216-9540-547eeaf67972\") " Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.567796 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-scripts" (OuterVolumeSpecName: "scripts") pod "4366459e-a3b8-4216-9540-547eeaf67972" (UID: "4366459e-a3b8-4216-9540-547eeaf67972"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.571359 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4366459e-a3b8-4216-9540-547eeaf67972-kube-api-access-66d8g" (OuterVolumeSpecName: "kube-api-access-66d8g") pod "4366459e-a3b8-4216-9540-547eeaf67972" (UID: "4366459e-a3b8-4216-9540-547eeaf67972"). InnerVolumeSpecName "kube-api-access-66d8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.593538 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4366459e-a3b8-4216-9540-547eeaf67972" (UID: "4366459e-a3b8-4216-9540-547eeaf67972"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.593665 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-config-data" (OuterVolumeSpecName: "config-data") pod "4366459e-a3b8-4216-9540-547eeaf67972" (UID: "4366459e-a3b8-4216-9540-547eeaf67972"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.665491 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.665548 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.665569 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66d8g\" (UniqueName: \"kubernetes.io/projected/4366459e-a3b8-4216-9540-547eeaf67972-kube-api-access-66d8g\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.665590 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4366459e-a3b8-4216-9540-547eeaf67972-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.922645 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vpbfr" Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.926072 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vpbfr" event={"ID":"4366459e-a3b8-4216-9540-547eeaf67972","Type":"ContainerDied","Data":"a50928e445f02d54bbc6da37bf92bd68bdb6251f8a1608256aeff3c310f7a3ae"} Nov 24 17:23:24 crc kubenswrapper[4777]: I1124 17:23:24.926126 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a50928e445f02d54bbc6da37bf92bd68bdb6251f8a1608256aeff3c310f7a3ae" Nov 24 17:23:25 crc kubenswrapper[4777]: I1124 17:23:25.113542 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:23:25 crc kubenswrapper[4777]: I1124 17:23:25.113817 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a7bceb11-3897-46ea-acd5-afeee23f55a1" containerName="nova-api-log" containerID="cri-o://3b269dd6b8f7d4e44841d81bf8334b719036d567161d55cad4818734cc18a246" gracePeriod=30 Nov 24 17:23:25 crc kubenswrapper[4777]: I1124 17:23:25.113917 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a7bceb11-3897-46ea-acd5-afeee23f55a1" containerName="nova-api-api" containerID="cri-o://959e2278072dd12b5d8f2c584e32b52da47ef209b072babfd7e5344a973e4cfe" gracePeriod=30 Nov 24 17:23:25 crc kubenswrapper[4777]: I1124 17:23:25.126948 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:23:25 crc kubenswrapper[4777]: I1124 17:23:25.127242 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="67dde531-8482-4ab8-9da0-b07999fde7fb" containerName="nova-scheduler-scheduler" containerID="cri-o://222cac5b4bc9425b5bf197e8939456f1fa7bf2cd2ad09ca3b679b9bd52f61842" gracePeriod=30 Nov 24 17:23:25 crc kubenswrapper[4777]: I1124 17:23:25.187465 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:23:25 crc kubenswrapper[4777]: I1124 17:23:25.650549 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-58bd69657f-bbqzr" podUID="2933da9d-6e63-4382-9e7c-6fc6437a7d1c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.190:5353: i/o timeout" Nov 24 17:23:25 crc kubenswrapper[4777]: I1124 17:23:25.971328 4777 generic.go:334] "Generic (PLEG): container finished" podID="a7bceb11-3897-46ea-acd5-afeee23f55a1" containerID="959e2278072dd12b5d8f2c584e32b52da47ef209b072babfd7e5344a973e4cfe" exitCode=0 Nov 24 17:23:25 crc kubenswrapper[4777]: I1124 17:23:25.971647 4777 generic.go:334] "Generic (PLEG): container finished" podID="a7bceb11-3897-46ea-acd5-afeee23f55a1" containerID="3b269dd6b8f7d4e44841d81bf8334b719036d567161d55cad4818734cc18a246" exitCode=143 Nov 24 17:23:25 crc kubenswrapper[4777]: I1124 17:23:25.971813 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="85940131-d100-4c1c-876e-3bea01dcc20b" containerName="nova-metadata-log" containerID="cri-o://82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555" gracePeriod=30 Nov 24 17:23:25 crc kubenswrapper[4777]: I1124 17:23:25.971388 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7bceb11-3897-46ea-acd5-afeee23f55a1","Type":"ContainerDied","Data":"959e2278072dd12b5d8f2c584e32b52da47ef209b072babfd7e5344a973e4cfe"} Nov 24 17:23:25 crc kubenswrapper[4777]: I1124 17:23:25.971895 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7bceb11-3897-46ea-acd5-afeee23f55a1","Type":"ContainerDied","Data":"3b269dd6b8f7d4e44841d81bf8334b719036d567161d55cad4818734cc18a246"} Nov 24 17:23:25 crc kubenswrapper[4777]: I1124 17:23:25.972108 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="85940131-d100-4c1c-876e-3bea01dcc20b" containerName="nova-metadata-metadata" containerID="cri-o://d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e" gracePeriod=30 Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.258601 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.405147 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7bceb11-3897-46ea-acd5-afeee23f55a1-logs\") pod \"a7bceb11-3897-46ea-acd5-afeee23f55a1\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.405522 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bceb11-3897-46ea-acd5-afeee23f55a1-combined-ca-bundle\") pod \"a7bceb11-3897-46ea-acd5-afeee23f55a1\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.405648 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bceb11-3897-46ea-acd5-afeee23f55a1-config-data\") pod \"a7bceb11-3897-46ea-acd5-afeee23f55a1\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.405762 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7wbm\" (UniqueName: \"kubernetes.io/projected/a7bceb11-3897-46ea-acd5-afeee23f55a1-kube-api-access-b7wbm\") pod \"a7bceb11-3897-46ea-acd5-afeee23f55a1\" (UID: \"a7bceb11-3897-46ea-acd5-afeee23f55a1\") " Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.406230 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7bceb11-3897-46ea-acd5-afeee23f55a1-logs" (OuterVolumeSpecName: "logs") pod "a7bceb11-3897-46ea-acd5-afeee23f55a1" (UID: "a7bceb11-3897-46ea-acd5-afeee23f55a1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.409945 4777 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7bceb11-3897-46ea-acd5-afeee23f55a1-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.434197 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7bceb11-3897-46ea-acd5-afeee23f55a1-kube-api-access-b7wbm" (OuterVolumeSpecName: "kube-api-access-b7wbm") pod "a7bceb11-3897-46ea-acd5-afeee23f55a1" (UID: "a7bceb11-3897-46ea-acd5-afeee23f55a1"). InnerVolumeSpecName "kube-api-access-b7wbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.442645 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7bceb11-3897-46ea-acd5-afeee23f55a1-config-data" (OuterVolumeSpecName: "config-data") pod "a7bceb11-3897-46ea-acd5-afeee23f55a1" (UID: "a7bceb11-3897-46ea-acd5-afeee23f55a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.458290 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7bceb11-3897-46ea-acd5-afeee23f55a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7bceb11-3897-46ea-acd5-afeee23f55a1" (UID: "a7bceb11-3897-46ea-acd5-afeee23f55a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.511961 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bceb11-3897-46ea-acd5-afeee23f55a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.512004 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bceb11-3897-46ea-acd5-afeee23f55a1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.512014 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7wbm\" (UniqueName: \"kubernetes.io/projected/a7bceb11-3897-46ea-acd5-afeee23f55a1-kube-api-access-b7wbm\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.589528 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.589591 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.652875 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.714703 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-config-data\") pod \"85940131-d100-4c1c-876e-3bea01dcc20b\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.714841 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85940131-d100-4c1c-876e-3bea01dcc20b-logs\") pod \"85940131-d100-4c1c-876e-3bea01dcc20b\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.714896 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-combined-ca-bundle\") pod \"85940131-d100-4c1c-876e-3bea01dcc20b\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.714954 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kqvq\" (UniqueName: \"kubernetes.io/projected/85940131-d100-4c1c-876e-3bea01dcc20b-kube-api-access-8kqvq\") pod \"85940131-d100-4c1c-876e-3bea01dcc20b\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.715125 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-nova-metadata-tls-certs\") pod \"85940131-d100-4c1c-876e-3bea01dcc20b\" (UID: \"85940131-d100-4c1c-876e-3bea01dcc20b\") " Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.716303 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85940131-d100-4c1c-876e-3bea01dcc20b-logs" (OuterVolumeSpecName: "logs") pod "85940131-d100-4c1c-876e-3bea01dcc20b" (UID: "85940131-d100-4c1c-876e-3bea01dcc20b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.739151 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85940131-d100-4c1c-876e-3bea01dcc20b-kube-api-access-8kqvq" (OuterVolumeSpecName: "kube-api-access-8kqvq") pod "85940131-d100-4c1c-876e-3bea01dcc20b" (UID: "85940131-d100-4c1c-876e-3bea01dcc20b"). InnerVolumeSpecName "kube-api-access-8kqvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.746187 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-config-data" (OuterVolumeSpecName: "config-data") pod "85940131-d100-4c1c-876e-3bea01dcc20b" (UID: "85940131-d100-4c1c-876e-3bea01dcc20b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.793261 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85940131-d100-4c1c-876e-3bea01dcc20b" (UID: "85940131-d100-4c1c-876e-3bea01dcc20b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.806283 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "85940131-d100-4c1c-876e-3bea01dcc20b" (UID: "85940131-d100-4c1c-876e-3bea01dcc20b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.819874 4777 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.820132 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.820206 4777 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85940131-d100-4c1c-876e-3bea01dcc20b-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.820268 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85940131-d100-4c1c-876e-3bea01dcc20b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.820338 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kqvq\" (UniqueName: \"kubernetes.io/projected/85940131-d100-4c1c-876e-3bea01dcc20b-kube-api-access-8kqvq\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.980880 4777 generic.go:334] "Generic (PLEG): container finished" podID="85940131-d100-4c1c-876e-3bea01dcc20b" containerID="d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e" exitCode=0 Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.980910 4777 generic.go:334] "Generic (PLEG): container finished" podID="85940131-d100-4c1c-876e-3bea01dcc20b" containerID="82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555" exitCode=143 Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.980955 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"85940131-d100-4c1c-876e-3bea01dcc20b","Type":"ContainerDied","Data":"d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e"} Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.980956 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.980999 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"85940131-d100-4c1c-876e-3bea01dcc20b","Type":"ContainerDied","Data":"82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555"} Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.981011 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"85940131-d100-4c1c-876e-3bea01dcc20b","Type":"ContainerDied","Data":"b5fbe4365e942f99258cdf99dadc77d311064f5835444ff19db8db02c97f9a38"} Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.981025 4777 scope.go:117] "RemoveContainer" containerID="d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e" Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.983187 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7bceb11-3897-46ea-acd5-afeee23f55a1","Type":"ContainerDied","Data":"3958b07eb74156bdf7e9b77b216b3e307e839a042bd1b4f657a3e363b5c0d031"} Nov 24 17:23:26 crc kubenswrapper[4777]: I1124 17:23:26.983250 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.014652 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.014945 4777 scope.go:117] "RemoveContainer" containerID="82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.024048 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.040789 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.046892 4777 scope.go:117] "RemoveContainer" containerID="d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e" Nov 24 17:23:27 crc kubenswrapper[4777]: E1124 17:23:27.047617 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e\": container with ID starting with d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e not found: ID does not exist" containerID="d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.047658 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e"} err="failed to get container status \"d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e\": rpc error: code = NotFound desc = could not find container \"d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e\": container with ID starting with d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e not found: ID does not exist" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.047686 4777 scope.go:117] "RemoveContainer" containerID="82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555" Nov 24 17:23:27 crc kubenswrapper[4777]: E1124 17:23:27.048490 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555\": container with ID starting with 82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555 not found: ID does not exist" containerID="82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.048509 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555"} err="failed to get container status \"82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555\": rpc error: code = NotFound desc = could not find container \"82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555\": container with ID starting with 82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555 not found: ID does not exist" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.048525 4777 scope.go:117] "RemoveContainer" containerID="d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.050344 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e"} err="failed to get container status \"d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e\": rpc error: code = NotFound desc = could not find container \"d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e\": container with ID starting with d2f3f0aa05dd6f989f85b6ce7d57de6d3822189c6954290720336d9ec9d7fe3e not found: ID does not exist" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.050366 4777 scope.go:117] "RemoveContainer" containerID="82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.050608 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555"} err="failed to get container status \"82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555\": rpc error: code = NotFound desc = could not find container \"82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555\": container with ID starting with 82dbc4626bf33a09e09d11ff0a83fc8cac5d519d2887e4369dc5b58fe6956555 not found: ID does not exist" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.050644 4777 scope.go:117] "RemoveContainer" containerID="959e2278072dd12b5d8f2c584e32b52da47ef209b072babfd7e5344a973e4cfe" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.055065 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.071724 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:23:27 crc kubenswrapper[4777]: E1124 17:23:27.072360 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7bceb11-3897-46ea-acd5-afeee23f55a1" containerName="nova-api-log" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.072377 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7bceb11-3897-46ea-acd5-afeee23f55a1" containerName="nova-api-log" Nov 24 17:23:27 crc kubenswrapper[4777]: E1124 17:23:27.072392 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85940131-d100-4c1c-876e-3bea01dcc20b" containerName="nova-metadata-metadata" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.072399 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="85940131-d100-4c1c-876e-3bea01dcc20b" containerName="nova-metadata-metadata" Nov 24 17:23:27 crc kubenswrapper[4777]: E1124 17:23:27.072438 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85940131-d100-4c1c-876e-3bea01dcc20b" containerName="nova-metadata-log" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.072444 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="85940131-d100-4c1c-876e-3bea01dcc20b" containerName="nova-metadata-log" Nov 24 17:23:27 crc kubenswrapper[4777]: E1124 17:23:27.072461 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4366459e-a3b8-4216-9540-547eeaf67972" containerName="nova-manage" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.072467 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="4366459e-a3b8-4216-9540-547eeaf67972" containerName="nova-manage" Nov 24 17:23:27 crc kubenswrapper[4777]: E1124 17:23:27.072479 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7bceb11-3897-46ea-acd5-afeee23f55a1" containerName="nova-api-api" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.072486 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7bceb11-3897-46ea-acd5-afeee23f55a1" containerName="nova-api-api" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.072660 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="85940131-d100-4c1c-876e-3bea01dcc20b" containerName="nova-metadata-log" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.072671 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7bceb11-3897-46ea-acd5-afeee23f55a1" containerName="nova-api-log" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.072683 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="4366459e-a3b8-4216-9540-547eeaf67972" containerName="nova-manage" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.072694 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7bceb11-3897-46ea-acd5-afeee23f55a1" containerName="nova-api-api" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.072711 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="85940131-d100-4c1c-876e-3bea01dcc20b" containerName="nova-metadata-metadata" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.073848 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.079178 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.079343 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.096942 4777 scope.go:117] "RemoveContainer" containerID="3b269dd6b8f7d4e44841d81bf8334b719036d567161d55cad4818734cc18a246" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.099886 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.116418 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.121368 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.124560 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.138680 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.236330 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.236415 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69de6629-9de1-41c1-aa41-4bb079a0505c-config-data\") pod \"nova-api-0\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " pod="openstack/nova-api-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.236499 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts725\" (UniqueName: \"kubernetes.io/projected/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-kube-api-access-ts725\") pod \"nova-metadata-0\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.236539 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-config-data\") pod \"nova-metadata-0\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.236587 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdl8x\" (UniqueName: \"kubernetes.io/projected/69de6629-9de1-41c1-aa41-4bb079a0505c-kube-api-access-tdl8x\") pod \"nova-api-0\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " pod="openstack/nova-api-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.236625 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69de6629-9de1-41c1-aa41-4bb079a0505c-logs\") pod \"nova-api-0\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " pod="openstack/nova-api-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.236678 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-logs\") pod \"nova-metadata-0\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.237065 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.245007 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69de6629-9de1-41c1-aa41-4bb079a0505c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " pod="openstack/nova-api-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.261215 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85940131-d100-4c1c-876e-3bea01dcc20b" path="/var/lib/kubelet/pods/85940131-d100-4c1c-876e-3bea01dcc20b/volumes" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.261798 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7bceb11-3897-46ea-acd5-afeee23f55a1" path="/var/lib/kubelet/pods/a7bceb11-3897-46ea-acd5-afeee23f55a1/volumes" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.348460 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69de6629-9de1-41c1-aa41-4bb079a0505c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " pod="openstack/nova-api-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.348519 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.348542 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69de6629-9de1-41c1-aa41-4bb079a0505c-config-data\") pod \"nova-api-0\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " pod="openstack/nova-api-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.348582 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts725\" (UniqueName: \"kubernetes.io/projected/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-kube-api-access-ts725\") pod \"nova-metadata-0\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.348604 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-config-data\") pod \"nova-metadata-0\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.348626 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdl8x\" (UniqueName: \"kubernetes.io/projected/69de6629-9de1-41c1-aa41-4bb079a0505c-kube-api-access-tdl8x\") pod \"nova-api-0\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " pod="openstack/nova-api-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.348650 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69de6629-9de1-41c1-aa41-4bb079a0505c-logs\") pod \"nova-api-0\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " pod="openstack/nova-api-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.348677 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-logs\") pod \"nova-metadata-0\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.348743 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.352936 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-logs\") pod \"nova-metadata-0\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.354128 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69de6629-9de1-41c1-aa41-4bb079a0505c-logs\") pod \"nova-api-0\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " pod="openstack/nova-api-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.354315 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69de6629-9de1-41c1-aa41-4bb079a0505c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " pod="openstack/nova-api-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.354646 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.355500 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-config-data\") pod \"nova-metadata-0\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.356760 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69de6629-9de1-41c1-aa41-4bb079a0505c-config-data\") pod \"nova-api-0\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " pod="openstack/nova-api-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.366877 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.373422 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdl8x\" (UniqueName: \"kubernetes.io/projected/69de6629-9de1-41c1-aa41-4bb079a0505c-kube-api-access-tdl8x\") pod \"nova-api-0\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " pod="openstack/nova-api-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.376130 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts725\" (UniqueName: \"kubernetes.io/projected/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-kube-api-access-ts725\") pod \"nova-metadata-0\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.442498 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.452584 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.647802 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.755488 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z8ln\" (UniqueName: \"kubernetes.io/projected/67dde531-8482-4ab8-9da0-b07999fde7fb-kube-api-access-9z8ln\") pod \"67dde531-8482-4ab8-9da0-b07999fde7fb\" (UID: \"67dde531-8482-4ab8-9da0-b07999fde7fb\") " Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.755620 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67dde531-8482-4ab8-9da0-b07999fde7fb-combined-ca-bundle\") pod \"67dde531-8482-4ab8-9da0-b07999fde7fb\" (UID: \"67dde531-8482-4ab8-9da0-b07999fde7fb\") " Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.755726 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67dde531-8482-4ab8-9da0-b07999fde7fb-config-data\") pod \"67dde531-8482-4ab8-9da0-b07999fde7fb\" (UID: \"67dde531-8482-4ab8-9da0-b07999fde7fb\") " Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.761467 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67dde531-8482-4ab8-9da0-b07999fde7fb-kube-api-access-9z8ln" (OuterVolumeSpecName: "kube-api-access-9z8ln") pod "67dde531-8482-4ab8-9da0-b07999fde7fb" (UID: "67dde531-8482-4ab8-9da0-b07999fde7fb"). InnerVolumeSpecName "kube-api-access-9z8ln". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.791824 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67dde531-8482-4ab8-9da0-b07999fde7fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67dde531-8482-4ab8-9da0-b07999fde7fb" (UID: "67dde531-8482-4ab8-9da0-b07999fde7fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.853310 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67dde531-8482-4ab8-9da0-b07999fde7fb-config-data" (OuterVolumeSpecName: "config-data") pod "67dde531-8482-4ab8-9da0-b07999fde7fb" (UID: "67dde531-8482-4ab8-9da0-b07999fde7fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.858175 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9z8ln\" (UniqueName: \"kubernetes.io/projected/67dde531-8482-4ab8-9da0-b07999fde7fb-kube-api-access-9z8ln\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.858205 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67dde531-8482-4ab8-9da0-b07999fde7fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.858220 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67dde531-8482-4ab8-9da0-b07999fde7fb-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.977103 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:23:27 crc kubenswrapper[4777]: I1124 17:23:27.999372 4777 generic.go:334] "Generic (PLEG): container finished" podID="67dde531-8482-4ab8-9da0-b07999fde7fb" containerID="222cac5b4bc9425b5bf197e8939456f1fa7bf2cd2ad09ca3b679b9bd52f61842" exitCode=0 Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:27.999431 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"67dde531-8482-4ab8-9da0-b07999fde7fb","Type":"ContainerDied","Data":"222cac5b4bc9425b5bf197e8939456f1fa7bf2cd2ad09ca3b679b9bd52f61842"} Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:27.999458 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"67dde531-8482-4ab8-9da0-b07999fde7fb","Type":"ContainerDied","Data":"80a6ccaa65db195566ee0fb259cba0c784db433d6d5e85a86707b9f93f3fee3d"} Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:27.999474 4777 scope.go:117] "RemoveContainer" containerID="222cac5b4bc9425b5bf197e8939456f1fa7bf2cd2ad09ca3b679b9bd52f61842" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:27.999572 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.002060 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf","Type":"ContainerStarted","Data":"ce9699c18e2d659d0691a72b36cb9b9fa1fc8016bb37120187f123eae898da76"} Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.046488 4777 scope.go:117] "RemoveContainer" containerID="222cac5b4bc9425b5bf197e8939456f1fa7bf2cd2ad09ca3b679b9bd52f61842" Nov 24 17:23:28 crc kubenswrapper[4777]: E1124 17:23:28.047159 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"222cac5b4bc9425b5bf197e8939456f1fa7bf2cd2ad09ca3b679b9bd52f61842\": container with ID starting with 222cac5b4bc9425b5bf197e8939456f1fa7bf2cd2ad09ca3b679b9bd52f61842 not found: ID does not exist" containerID="222cac5b4bc9425b5bf197e8939456f1fa7bf2cd2ad09ca3b679b9bd52f61842" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.047212 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"222cac5b4bc9425b5bf197e8939456f1fa7bf2cd2ad09ca3b679b9bd52f61842"} err="failed to get container status \"222cac5b4bc9425b5bf197e8939456f1fa7bf2cd2ad09ca3b679b9bd52f61842\": rpc error: code = NotFound desc = could not find container \"222cac5b4bc9425b5bf197e8939456f1fa7bf2cd2ad09ca3b679b9bd52f61842\": container with ID starting with 222cac5b4bc9425b5bf197e8939456f1fa7bf2cd2ad09ca3b679b9bd52f61842 not found: ID does not exist" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.049338 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.097549 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.113269 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:23:28 crc kubenswrapper[4777]: E1124 17:23:28.113687 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67dde531-8482-4ab8-9da0-b07999fde7fb" containerName="nova-scheduler-scheduler" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.113703 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="67dde531-8482-4ab8-9da0-b07999fde7fb" containerName="nova-scheduler-scheduler" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.115620 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="67dde531-8482-4ab8-9da0-b07999fde7fb" containerName="nova-scheduler-scheduler" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.116544 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.119803 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.124319 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.163387 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:23:28 crc kubenswrapper[4777]: W1124 17:23:28.164385 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69de6629_9de1_41c1_aa41_4bb079a0505c.slice/crio-2cce1bc7fd0c41bc72f125a2975c1eda218a1a79194f68b899aae81aa979036f WatchSource:0}: Error finding container 2cce1bc7fd0c41bc72f125a2975c1eda218a1a79194f68b899aae81aa979036f: Status 404 returned error can't find the container with id 2cce1bc7fd0c41bc72f125a2975c1eda218a1a79194f68b899aae81aa979036f Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.267525 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9b3573-39b1-4336-9048-670f2ba47c23-config-data\") pod \"nova-scheduler-0\" (UID: \"5a9b3573-39b1-4336-9048-670f2ba47c23\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.267644 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9b3573-39b1-4336-9048-670f2ba47c23-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5a9b3573-39b1-4336-9048-670f2ba47c23\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.267677 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bldlw\" (UniqueName: \"kubernetes.io/projected/5a9b3573-39b1-4336-9048-670f2ba47c23-kube-api-access-bldlw\") pod \"nova-scheduler-0\" (UID: \"5a9b3573-39b1-4336-9048-670f2ba47c23\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.369218 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9b3573-39b1-4336-9048-670f2ba47c23-config-data\") pod \"nova-scheduler-0\" (UID: \"5a9b3573-39b1-4336-9048-670f2ba47c23\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.369312 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9b3573-39b1-4336-9048-670f2ba47c23-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5a9b3573-39b1-4336-9048-670f2ba47c23\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.369346 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bldlw\" (UniqueName: \"kubernetes.io/projected/5a9b3573-39b1-4336-9048-670f2ba47c23-kube-api-access-bldlw\") pod \"nova-scheduler-0\" (UID: \"5a9b3573-39b1-4336-9048-670f2ba47c23\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.373031 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9b3573-39b1-4336-9048-670f2ba47c23-config-data\") pod \"nova-scheduler-0\" (UID: \"5a9b3573-39b1-4336-9048-670f2ba47c23\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.374257 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9b3573-39b1-4336-9048-670f2ba47c23-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5a9b3573-39b1-4336-9048-670f2ba47c23\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.394874 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bldlw\" (UniqueName: \"kubernetes.io/projected/5a9b3573-39b1-4336-9048-670f2ba47c23-kube-api-access-bldlw\") pod \"nova-scheduler-0\" (UID: \"5a9b3573-39b1-4336-9048-670f2ba47c23\") " pod="openstack/nova-scheduler-0" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.438760 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:23:28 crc kubenswrapper[4777]: I1124 17:23:28.960040 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:23:29 crc kubenswrapper[4777]: I1124 17:23:29.023399 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf","Type":"ContainerStarted","Data":"ad8e30e2fb3ae38dd466b51ed874992a582c1b29e9ef0571df393fd62f4276ea"} Nov 24 17:23:29 crc kubenswrapper[4777]: I1124 17:23:29.023444 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf","Type":"ContainerStarted","Data":"f04868a0ddb0b3bf2ae3d10e99a7f026574673b7cf2a17a00218da9b22fa4e23"} Nov 24 17:23:29 crc kubenswrapper[4777]: I1124 17:23:29.029878 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"69de6629-9de1-41c1-aa41-4bb079a0505c","Type":"ContainerStarted","Data":"fc4e8352af11ae0b15300cb0dc91ff08b6acc39f8ebf5948877c6b153e61dbc7"} Nov 24 17:23:29 crc kubenswrapper[4777]: I1124 17:23:29.029943 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"69de6629-9de1-41c1-aa41-4bb079a0505c","Type":"ContainerStarted","Data":"86b79550368087fc8250f42d992d4f72846badbbc81795a5d351e4c7fb82b53d"} Nov 24 17:23:29 crc kubenswrapper[4777]: I1124 17:23:29.029959 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"69de6629-9de1-41c1-aa41-4bb079a0505c","Type":"ContainerStarted","Data":"2cce1bc7fd0c41bc72f125a2975c1eda218a1a79194f68b899aae81aa979036f"} Nov 24 17:23:29 crc kubenswrapper[4777]: I1124 17:23:29.032032 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a9b3573-39b1-4336-9048-670f2ba47c23","Type":"ContainerStarted","Data":"e31180a6f104245a6d82773a50030c76e48fdb34ae48d193a79cacce67d35746"} Nov 24 17:23:29 crc kubenswrapper[4777]: I1124 17:23:29.055345 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.05532622 podStartE2EDuration="2.05532622s" podCreationTimestamp="2025-11-24 17:23:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:23:29.04010464 +0000 UTC m=+1367.199299689" watchObservedRunningTime="2025-11-24 17:23:29.05532622 +0000 UTC m=+1367.214521269" Nov 24 17:23:29 crc kubenswrapper[4777]: I1124 17:23:29.072111 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.072088125 podStartE2EDuration="2.072088125s" podCreationTimestamp="2025-11-24 17:23:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:23:29.062062385 +0000 UTC m=+1367.221257434" watchObservedRunningTime="2025-11-24 17:23:29.072088125 +0000 UTC m=+1367.231283174" Nov 24 17:23:29 crc kubenswrapper[4777]: I1124 17:23:29.256577 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67dde531-8482-4ab8-9da0-b07999fde7fb" path="/var/lib/kubelet/pods/67dde531-8482-4ab8-9da0-b07999fde7fb/volumes" Nov 24 17:23:30 crc kubenswrapper[4777]: I1124 17:23:30.046098 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a9b3573-39b1-4336-9048-670f2ba47c23","Type":"ContainerStarted","Data":"24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c"} Nov 24 17:23:30 crc kubenswrapper[4777]: I1124 17:23:30.048337 4777 generic.go:334] "Generic (PLEG): container finished" podID="7d534624-b898-4268-a432-7f7a575fb15b" containerID="2c0a4fc5e29a49f0e4e2d0e4b5ffd7824db5cf109cd192786408a0ac53f33e7a" exitCode=0 Nov 24 17:23:30 crc kubenswrapper[4777]: I1124 17:23:30.048426 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xfksm" event={"ID":"7d534624-b898-4268-a432-7f7a575fb15b","Type":"ContainerDied","Data":"2c0a4fc5e29a49f0e4e2d0e4b5ffd7824db5cf109cd192786408a0ac53f33e7a"} Nov 24 17:23:30 crc kubenswrapper[4777]: I1124 17:23:30.075085 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.075055218 podStartE2EDuration="2.075055218s" podCreationTimestamp="2025-11-24 17:23:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:23:30.064398319 +0000 UTC m=+1368.223593408" watchObservedRunningTime="2025-11-24 17:23:30.075055218 +0000 UTC m=+1368.234250307" Nov 24 17:23:31 crc kubenswrapper[4777]: I1124 17:23:31.581470 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:31 crc kubenswrapper[4777]: I1124 17:23:31.740471 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-config-data\") pod \"7d534624-b898-4268-a432-7f7a575fb15b\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " Nov 24 17:23:31 crc kubenswrapper[4777]: I1124 17:23:31.740699 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-scripts\") pod \"7d534624-b898-4268-a432-7f7a575fb15b\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " Nov 24 17:23:31 crc kubenswrapper[4777]: I1124 17:23:31.740771 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9dwg\" (UniqueName: \"kubernetes.io/projected/7d534624-b898-4268-a432-7f7a575fb15b-kube-api-access-p9dwg\") pod \"7d534624-b898-4268-a432-7f7a575fb15b\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " Nov 24 17:23:31 crc kubenswrapper[4777]: I1124 17:23:31.740826 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-combined-ca-bundle\") pod \"7d534624-b898-4268-a432-7f7a575fb15b\" (UID: \"7d534624-b898-4268-a432-7f7a575fb15b\") " Nov 24 17:23:31 crc kubenswrapper[4777]: I1124 17:23:31.746439 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-scripts" (OuterVolumeSpecName: "scripts") pod "7d534624-b898-4268-a432-7f7a575fb15b" (UID: "7d534624-b898-4268-a432-7f7a575fb15b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:31 crc kubenswrapper[4777]: I1124 17:23:31.751112 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d534624-b898-4268-a432-7f7a575fb15b-kube-api-access-p9dwg" (OuterVolumeSpecName: "kube-api-access-p9dwg") pod "7d534624-b898-4268-a432-7f7a575fb15b" (UID: "7d534624-b898-4268-a432-7f7a575fb15b"). InnerVolumeSpecName "kube-api-access-p9dwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:23:31 crc kubenswrapper[4777]: I1124 17:23:31.783854 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-config-data" (OuterVolumeSpecName: "config-data") pod "7d534624-b898-4268-a432-7f7a575fb15b" (UID: "7d534624-b898-4268-a432-7f7a575fb15b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:31 crc kubenswrapper[4777]: I1124 17:23:31.784652 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d534624-b898-4268-a432-7f7a575fb15b" (UID: "7d534624-b898-4268-a432-7f7a575fb15b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:31 crc kubenswrapper[4777]: I1124 17:23:31.842723 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:31 crc kubenswrapper[4777]: I1124 17:23:31.842756 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9dwg\" (UniqueName: \"kubernetes.io/projected/7d534624-b898-4268-a432-7f7a575fb15b-kube-api-access-p9dwg\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:31 crc kubenswrapper[4777]: I1124 17:23:31.842770 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:31 crc kubenswrapper[4777]: I1124 17:23:31.842783 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d534624-b898-4268-a432-7f7a575fb15b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.072393 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xfksm" event={"ID":"7d534624-b898-4268-a432-7f7a575fb15b","Type":"ContainerDied","Data":"702d6e67650ae256ab18f6f334c074ea12f91c91becfecdc7b4b561d6d3b3bf9"} Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.072423 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xfksm" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.072435 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="702d6e67650ae256ab18f6f334c074ea12f91c91becfecdc7b4b561d6d3b3bf9" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.157766 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 17:23:32 crc kubenswrapper[4777]: E1124 17:23:32.158280 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d534624-b898-4268-a432-7f7a575fb15b" containerName="nova-cell1-conductor-db-sync" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.158304 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d534624-b898-4268-a432-7f7a575fb15b" containerName="nova-cell1-conductor-db-sync" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.158592 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d534624-b898-4268-a432-7f7a575fb15b" containerName="nova-cell1-conductor-db-sync" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.159560 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.162275 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.173447 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.252846 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b929fcfa-b979-4f40-a469-fdd1e24a9ef2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b929fcfa-b979-4f40-a469-fdd1e24a9ef2\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.252903 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xrkb\" (UniqueName: \"kubernetes.io/projected/b929fcfa-b979-4f40-a469-fdd1e24a9ef2-kube-api-access-2xrkb\") pod \"nova-cell1-conductor-0\" (UID: \"b929fcfa-b979-4f40-a469-fdd1e24a9ef2\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.252977 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b929fcfa-b979-4f40-a469-fdd1e24a9ef2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b929fcfa-b979-4f40-a469-fdd1e24a9ef2\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.354878 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b929fcfa-b979-4f40-a469-fdd1e24a9ef2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b929fcfa-b979-4f40-a469-fdd1e24a9ef2\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.355189 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xrkb\" (UniqueName: \"kubernetes.io/projected/b929fcfa-b979-4f40-a469-fdd1e24a9ef2-kube-api-access-2xrkb\") pod \"nova-cell1-conductor-0\" (UID: \"b929fcfa-b979-4f40-a469-fdd1e24a9ef2\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.355236 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b929fcfa-b979-4f40-a469-fdd1e24a9ef2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b929fcfa-b979-4f40-a469-fdd1e24a9ef2\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.361637 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b929fcfa-b979-4f40-a469-fdd1e24a9ef2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b929fcfa-b979-4f40-a469-fdd1e24a9ef2\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.362048 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b929fcfa-b979-4f40-a469-fdd1e24a9ef2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b929fcfa-b979-4f40-a469-fdd1e24a9ef2\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.380183 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xrkb\" (UniqueName: \"kubernetes.io/projected/b929fcfa-b979-4f40-a469-fdd1e24a9ef2-kube-api-access-2xrkb\") pod \"nova-cell1-conductor-0\" (UID: \"b929fcfa-b979-4f40-a469-fdd1e24a9ef2\") " pod="openstack/nova-cell1-conductor-0" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.443014 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.443358 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 17:23:32 crc kubenswrapper[4777]: I1124 17:23:32.478729 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 17:23:33 crc kubenswrapper[4777]: I1124 17:23:33.096930 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 17:23:33 crc kubenswrapper[4777]: I1124 17:23:33.439559 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 17:23:34 crc kubenswrapper[4777]: I1124 17:23:34.107380 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b929fcfa-b979-4f40-a469-fdd1e24a9ef2","Type":"ContainerStarted","Data":"ffba818e09dd65db22f7b17819846aa44e61afe80303dc2d5eefaa7d863fc06d"} Nov 24 17:23:34 crc kubenswrapper[4777]: I1124 17:23:34.107444 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b929fcfa-b979-4f40-a469-fdd1e24a9ef2","Type":"ContainerStarted","Data":"d56a5a2d6969c30fb56bf003442d06dc92627342c90676d59a0d331bf37aa7f5"} Nov 24 17:23:34 crc kubenswrapper[4777]: I1124 17:23:34.108099 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 17:23:34 crc kubenswrapper[4777]: I1124 17:23:34.129495 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.129452492 podStartE2EDuration="2.129452492s" podCreationTimestamp="2025-11-24 17:23:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:23:34.125314143 +0000 UTC m=+1372.284509202" watchObservedRunningTime="2025-11-24 17:23:34.129452492 +0000 UTC m=+1372.288669202" Nov 24 17:23:37 crc kubenswrapper[4777]: I1124 17:23:37.443597 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 17:23:37 crc kubenswrapper[4777]: I1124 17:23:37.444161 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 17:23:37 crc kubenswrapper[4777]: I1124 17:23:37.453858 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 17:23:37 crc kubenswrapper[4777]: I1124 17:23:37.453914 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 17:23:38 crc kubenswrapper[4777]: I1124 17:23:38.439655 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 17:23:38 crc kubenswrapper[4777]: I1124 17:23:38.457266 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.217:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 17:23:38 crc kubenswrapper[4777]: I1124 17:23:38.457299 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.217:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 17:23:38 crc kubenswrapper[4777]: I1124 17:23:38.497626 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 17:23:38 crc kubenswrapper[4777]: I1124 17:23:38.541214 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="69de6629-9de1-41c1-aa41-4bb079a0505c" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.218:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 17:23:38 crc kubenswrapper[4777]: I1124 17:23:38.541220 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="69de6629-9de1-41c1-aa41-4bb079a0505c" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.218:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 17:23:39 crc kubenswrapper[4777]: I1124 17:23:39.201415 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 17:23:40 crc kubenswrapper[4777]: I1124 17:23:40.279661 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 17:23:42 crc kubenswrapper[4777]: I1124 17:23:42.525503 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 17:23:43 crc kubenswrapper[4777]: I1124 17:23:43.802773 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:23:43 crc kubenswrapper[4777]: I1124 17:23:43.803556 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="155b4504-c156-4af8-a9e0-1dc66c9d7d43" containerName="kube-state-metrics" containerID="cri-o://f46b0fb98030f89ed777d3807dcbac2a515d8a96488e3f477fe4aa40b840f5c6" gracePeriod=30 Nov 24 17:23:44 crc kubenswrapper[4777]: I1124 17:23:44.225802 4777 generic.go:334] "Generic (PLEG): container finished" podID="155b4504-c156-4af8-a9e0-1dc66c9d7d43" containerID="f46b0fb98030f89ed777d3807dcbac2a515d8a96488e3f477fe4aa40b840f5c6" exitCode=2 Nov 24 17:23:44 crc kubenswrapper[4777]: I1124 17:23:44.225858 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"155b4504-c156-4af8-a9e0-1dc66c9d7d43","Type":"ContainerDied","Data":"f46b0fb98030f89ed777d3807dcbac2a515d8a96488e3f477fe4aa40b840f5c6"} Nov 24 17:23:44 crc kubenswrapper[4777]: I1124 17:23:44.426151 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 17:23:44 crc kubenswrapper[4777]: I1124 17:23:44.514209 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9k48\" (UniqueName: \"kubernetes.io/projected/155b4504-c156-4af8-a9e0-1dc66c9d7d43-kube-api-access-x9k48\") pod \"155b4504-c156-4af8-a9e0-1dc66c9d7d43\" (UID: \"155b4504-c156-4af8-a9e0-1dc66c9d7d43\") " Nov 24 17:23:44 crc kubenswrapper[4777]: I1124 17:23:44.537219 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/155b4504-c156-4af8-a9e0-1dc66c9d7d43-kube-api-access-x9k48" (OuterVolumeSpecName: "kube-api-access-x9k48") pod "155b4504-c156-4af8-a9e0-1dc66c9d7d43" (UID: "155b4504-c156-4af8-a9e0-1dc66c9d7d43"). InnerVolumeSpecName "kube-api-access-x9k48". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:23:44 crc kubenswrapper[4777]: I1124 17:23:44.616449 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9k48\" (UniqueName: \"kubernetes.io/projected/155b4504-c156-4af8-a9e0-1dc66c9d7d43-kube-api-access-x9k48\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.247305 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.257581 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"155b4504-c156-4af8-a9e0-1dc66c9d7d43","Type":"ContainerDied","Data":"a97fdefbd5723c9e2d7d6333a32e3b2b7926b259d827a2b1b9e38959f41722bd"} Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.257631 4777 scope.go:117] "RemoveContainer" containerID="f46b0fb98030f89ed777d3807dcbac2a515d8a96488e3f477fe4aa40b840f5c6" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.295308 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.311114 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.323949 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:23:45 crc kubenswrapper[4777]: E1124 17:23:45.324444 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="155b4504-c156-4af8-a9e0-1dc66c9d7d43" containerName="kube-state-metrics" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.324463 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="155b4504-c156-4af8-a9e0-1dc66c9d7d43" containerName="kube-state-metrics" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.324673 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="155b4504-c156-4af8-a9e0-1dc66c9d7d43" containerName="kube-state-metrics" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.325467 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.334478 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.334731 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.340426 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.439579 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e1ab5a6-3bb5-4722-9220-1a033a656b93-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1e1ab5a6-3bb5-4722-9220-1a033a656b93\") " pod="openstack/kube-state-metrics-0" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.439632 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hh2j\" (UniqueName: \"kubernetes.io/projected/1e1ab5a6-3bb5-4722-9220-1a033a656b93-kube-api-access-6hh2j\") pod \"kube-state-metrics-0\" (UID: \"1e1ab5a6-3bb5-4722-9220-1a033a656b93\") " pod="openstack/kube-state-metrics-0" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.439707 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1e1ab5a6-3bb5-4722-9220-1a033a656b93-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1e1ab5a6-3bb5-4722-9220-1a033a656b93\") " pod="openstack/kube-state-metrics-0" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.439744 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e1ab5a6-3bb5-4722-9220-1a033a656b93-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1e1ab5a6-3bb5-4722-9220-1a033a656b93\") " pod="openstack/kube-state-metrics-0" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.541360 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e1ab5a6-3bb5-4722-9220-1a033a656b93-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1e1ab5a6-3bb5-4722-9220-1a033a656b93\") " pod="openstack/kube-state-metrics-0" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.541606 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hh2j\" (UniqueName: \"kubernetes.io/projected/1e1ab5a6-3bb5-4722-9220-1a033a656b93-kube-api-access-6hh2j\") pod \"kube-state-metrics-0\" (UID: \"1e1ab5a6-3bb5-4722-9220-1a033a656b93\") " pod="openstack/kube-state-metrics-0" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.541767 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1e1ab5a6-3bb5-4722-9220-1a033a656b93-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1e1ab5a6-3bb5-4722-9220-1a033a656b93\") " pod="openstack/kube-state-metrics-0" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.541856 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e1ab5a6-3bb5-4722-9220-1a033a656b93-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1e1ab5a6-3bb5-4722-9220-1a033a656b93\") " pod="openstack/kube-state-metrics-0" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.547570 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1e1ab5a6-3bb5-4722-9220-1a033a656b93-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1e1ab5a6-3bb5-4722-9220-1a033a656b93\") " pod="openstack/kube-state-metrics-0" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.550783 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e1ab5a6-3bb5-4722-9220-1a033a656b93-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1e1ab5a6-3bb5-4722-9220-1a033a656b93\") " pod="openstack/kube-state-metrics-0" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.554623 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e1ab5a6-3bb5-4722-9220-1a033a656b93-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1e1ab5a6-3bb5-4722-9220-1a033a656b93\") " pod="openstack/kube-state-metrics-0" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.562321 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hh2j\" (UniqueName: \"kubernetes.io/projected/1e1ab5a6-3bb5-4722-9220-1a033a656b93-kube-api-access-6hh2j\") pod \"kube-state-metrics-0\" (UID: \"1e1ab5a6-3bb5-4722-9220-1a033a656b93\") " pod="openstack/kube-state-metrics-0" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.663160 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.740085 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.740446 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="ceilometer-central-agent" containerID="cri-o://6e14cb85635d3e244139a9e247c2bce46931d36cabc3d9bd13250d204420dda6" gracePeriod=30 Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.741056 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="proxy-httpd" containerID="cri-o://6f8ee2273c0cc1a79dea4f80589dfbbdb693bba486bfd47d18b0c39f0810f381" gracePeriod=30 Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.741129 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="sg-core" containerID="cri-o://29d2470b7cadd938824606893257bf5cda27a59d52c06dea02db705e7e6d80ba" gracePeriod=30 Nov 24 17:23:45 crc kubenswrapper[4777]: I1124 17:23:45.741179 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="ceilometer-notification-agent" containerID="cri-o://c6aed0abf3b15443476bb3d78d14298f34006b46dd8b8b47d5549d7272fb324c" gracePeriod=30 Nov 24 17:23:46 crc kubenswrapper[4777]: I1124 17:23:46.174176 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 17:23:46 crc kubenswrapper[4777]: I1124 17:23:46.260073 4777 generic.go:334] "Generic (PLEG): container finished" podID="6882541d-9663-41db-9d0a-ec16ead4f421" containerID="6f8ee2273c0cc1a79dea4f80589dfbbdb693bba486bfd47d18b0c39f0810f381" exitCode=0 Nov 24 17:23:46 crc kubenswrapper[4777]: I1124 17:23:46.260103 4777 generic.go:334] "Generic (PLEG): container finished" podID="6882541d-9663-41db-9d0a-ec16ead4f421" containerID="29d2470b7cadd938824606893257bf5cda27a59d52c06dea02db705e7e6d80ba" exitCode=2 Nov 24 17:23:46 crc kubenswrapper[4777]: I1124 17:23:46.260112 4777 generic.go:334] "Generic (PLEG): container finished" podID="6882541d-9663-41db-9d0a-ec16ead4f421" containerID="6e14cb85635d3e244139a9e247c2bce46931d36cabc3d9bd13250d204420dda6" exitCode=0 Nov 24 17:23:46 crc kubenswrapper[4777]: I1124 17:23:46.260141 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6882541d-9663-41db-9d0a-ec16ead4f421","Type":"ContainerDied","Data":"6f8ee2273c0cc1a79dea4f80589dfbbdb693bba486bfd47d18b0c39f0810f381"} Nov 24 17:23:46 crc kubenswrapper[4777]: I1124 17:23:46.260177 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6882541d-9663-41db-9d0a-ec16ead4f421","Type":"ContainerDied","Data":"29d2470b7cadd938824606893257bf5cda27a59d52c06dea02db705e7e6d80ba"} Nov 24 17:23:46 crc kubenswrapper[4777]: I1124 17:23:46.260186 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6882541d-9663-41db-9d0a-ec16ead4f421","Type":"ContainerDied","Data":"6e14cb85635d3e244139a9e247c2bce46931d36cabc3d9bd13250d204420dda6"} Nov 24 17:23:46 crc kubenswrapper[4777]: I1124 17:23:46.262987 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1e1ab5a6-3bb5-4722-9220-1a033a656b93","Type":"ContainerStarted","Data":"ab8fabc9bf3d1d42fe4dffd496fd29d886a21df7fe3b5e3358976769eed2420b"} Nov 24 17:23:47 crc kubenswrapper[4777]: I1124 17:23:47.260486 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="155b4504-c156-4af8-a9e0-1dc66c9d7d43" path="/var/lib/kubelet/pods/155b4504-c156-4af8-a9e0-1dc66c9d7d43/volumes" Nov 24 17:23:47 crc kubenswrapper[4777]: I1124 17:23:47.278049 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1e1ab5a6-3bb5-4722-9220-1a033a656b93","Type":"ContainerStarted","Data":"af87694557ee8cdfb694ae803d04117a8b6bfaaea0cb8cf508992691125612a0"} Nov 24 17:23:47 crc kubenswrapper[4777]: I1124 17:23:47.278440 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 17:23:47 crc kubenswrapper[4777]: I1124 17:23:47.300417 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.920062599 podStartE2EDuration="2.300392668s" podCreationTimestamp="2025-11-24 17:23:45 +0000 UTC" firstStartedPulling="2025-11-24 17:23:46.186776565 +0000 UTC m=+1384.345971614" lastFinishedPulling="2025-11-24 17:23:46.567106634 +0000 UTC m=+1384.726301683" observedRunningTime="2025-11-24 17:23:47.293758496 +0000 UTC m=+1385.452953565" watchObservedRunningTime="2025-11-24 17:23:47.300392668 +0000 UTC m=+1385.459587727" Nov 24 17:23:47 crc kubenswrapper[4777]: I1124 17:23:47.447435 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 17:23:47 crc kubenswrapper[4777]: I1124 17:23:47.449932 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 17:23:47 crc kubenswrapper[4777]: I1124 17:23:47.468289 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 17:23:47 crc kubenswrapper[4777]: I1124 17:23:47.468664 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 17:23:47 crc kubenswrapper[4777]: I1124 17:23:47.472053 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 17:23:47 crc kubenswrapper[4777]: I1124 17:23:47.475525 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 17:23:47 crc kubenswrapper[4777]: I1124 17:23:47.477339 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.287558 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.291536 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.293126 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.480543 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54dd998c-8smx5"] Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.482828 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.493584 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54dd998c-8smx5"] Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.605442 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-dns-svc\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.605506 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-dns-swift-storage-0\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.605529 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-ovsdbserver-nb\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.605583 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-config\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.605629 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-ovsdbserver-sb\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.605657 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddht8\" (UniqueName: \"kubernetes.io/projected/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-kube-api-access-ddht8\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.707094 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-dns-svc\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.707180 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-dns-swift-storage-0\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.707205 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-ovsdbserver-nb\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.707253 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-config\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.707302 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-ovsdbserver-sb\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.707326 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddht8\" (UniqueName: \"kubernetes.io/projected/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-kube-api-access-ddht8\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.708222 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-dns-svc\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.708334 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-ovsdbserver-nb\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.708374 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-config\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.708778 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-ovsdbserver-sb\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.708947 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-dns-swift-storage-0\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.730774 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddht8\" (UniqueName: \"kubernetes.io/projected/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-kube-api-access-ddht8\") pod \"dnsmasq-dns-54dd998c-8smx5\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:48 crc kubenswrapper[4777]: I1124 17:23:48.836924 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:49 crc kubenswrapper[4777]: I1124 17:23:49.301625 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54dd998c-8smx5"] Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.309920 4777 generic.go:334] "Generic (PLEG): container finished" podID="67ecd2f1-6485-4d94-b66c-0bc0465ea695" containerID="f58ae0ff9d205c07f2d8d7a420739188b36a3777636f74796eec5753453b8abd" exitCode=137 Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.310005 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"67ecd2f1-6485-4d94-b66c-0bc0465ea695","Type":"ContainerDied","Data":"f58ae0ff9d205c07f2d8d7a420739188b36a3777636f74796eec5753453b8abd"} Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.311220 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"67ecd2f1-6485-4d94-b66c-0bc0465ea695","Type":"ContainerDied","Data":"207c191a791bd18fd51972fd1fc5ef64c31a1d8e3c2d6732710c44002b8498db"} Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.311290 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="207c191a791bd18fd51972fd1fc5ef64c31a1d8e3c2d6732710c44002b8498db" Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.312882 4777 generic.go:334] "Generic (PLEG): container finished" podID="4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4" containerID="8c327a473dbd138850bbd3af0d40527f3cba217e1499d8a8bc72b9e6602db3ec" exitCode=0 Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.313002 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54dd998c-8smx5" event={"ID":"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4","Type":"ContainerDied","Data":"8c327a473dbd138850bbd3af0d40527f3cba217e1499d8a8bc72b9e6602db3ec"} Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.313064 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54dd998c-8smx5" event={"ID":"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4","Type":"ContainerStarted","Data":"eb78c2d9a75d4d54545968a46cc226dcf857934c931503744f2920b95884b23e"} Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.474576 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.544146 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ecd2f1-6485-4d94-b66c-0bc0465ea695-config-data\") pod \"67ecd2f1-6485-4d94-b66c-0bc0465ea695\" (UID: \"67ecd2f1-6485-4d94-b66c-0bc0465ea695\") " Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.544707 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk8f8\" (UniqueName: \"kubernetes.io/projected/67ecd2f1-6485-4d94-b66c-0bc0465ea695-kube-api-access-jk8f8\") pod \"67ecd2f1-6485-4d94-b66c-0bc0465ea695\" (UID: \"67ecd2f1-6485-4d94-b66c-0bc0465ea695\") " Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.544877 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ecd2f1-6485-4d94-b66c-0bc0465ea695-combined-ca-bundle\") pod \"67ecd2f1-6485-4d94-b66c-0bc0465ea695\" (UID: \"67ecd2f1-6485-4d94-b66c-0bc0465ea695\") " Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.555981 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67ecd2f1-6485-4d94-b66c-0bc0465ea695-kube-api-access-jk8f8" (OuterVolumeSpecName: "kube-api-access-jk8f8") pod "67ecd2f1-6485-4d94-b66c-0bc0465ea695" (UID: "67ecd2f1-6485-4d94-b66c-0bc0465ea695"). InnerVolumeSpecName "kube-api-access-jk8f8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.600161 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67ecd2f1-6485-4d94-b66c-0bc0465ea695-config-data" (OuterVolumeSpecName: "config-data") pod "67ecd2f1-6485-4d94-b66c-0bc0465ea695" (UID: "67ecd2f1-6485-4d94-b66c-0bc0465ea695"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.600195 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67ecd2f1-6485-4d94-b66c-0bc0465ea695-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67ecd2f1-6485-4d94-b66c-0bc0465ea695" (UID: "67ecd2f1-6485-4d94-b66c-0bc0465ea695"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.647771 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ecd2f1-6485-4d94-b66c-0bc0465ea695-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.647799 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk8f8\" (UniqueName: \"kubernetes.io/projected/67ecd2f1-6485-4d94-b66c-0bc0465ea695-kube-api-access-jk8f8\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:50 crc kubenswrapper[4777]: I1124 17:23:50.647810 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ecd2f1-6485-4d94-b66c-0bc0465ea695-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.176400 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.266565 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-scripts\") pod \"6882541d-9663-41db-9d0a-ec16ead4f421\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.266602 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-sg-core-conf-yaml\") pod \"6882541d-9663-41db-9d0a-ec16ead4f421\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.266636 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6882541d-9663-41db-9d0a-ec16ead4f421-run-httpd\") pod \"6882541d-9663-41db-9d0a-ec16ead4f421\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.266724 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6882541d-9663-41db-9d0a-ec16ead4f421-log-httpd\") pod \"6882541d-9663-41db-9d0a-ec16ead4f421\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.266791 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vcb4\" (UniqueName: \"kubernetes.io/projected/6882541d-9663-41db-9d0a-ec16ead4f421-kube-api-access-4vcb4\") pod \"6882541d-9663-41db-9d0a-ec16ead4f421\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.266843 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-config-data\") pod \"6882541d-9663-41db-9d0a-ec16ead4f421\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.266891 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-combined-ca-bundle\") pod \"6882541d-9663-41db-9d0a-ec16ead4f421\" (UID: \"6882541d-9663-41db-9d0a-ec16ead4f421\") " Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.267295 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6882541d-9663-41db-9d0a-ec16ead4f421-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6882541d-9663-41db-9d0a-ec16ead4f421" (UID: "6882541d-9663-41db-9d0a-ec16ead4f421"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.267752 4777 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6882541d-9663-41db-9d0a-ec16ead4f421-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.269352 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6882541d-9663-41db-9d0a-ec16ead4f421-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6882541d-9663-41db-9d0a-ec16ead4f421" (UID: "6882541d-9663-41db-9d0a-ec16ead4f421"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.272031 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-scripts" (OuterVolumeSpecName: "scripts") pod "6882541d-9663-41db-9d0a-ec16ead4f421" (UID: "6882541d-9663-41db-9d0a-ec16ead4f421"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.280445 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6882541d-9663-41db-9d0a-ec16ead4f421-kube-api-access-4vcb4" (OuterVolumeSpecName: "kube-api-access-4vcb4") pod "6882541d-9663-41db-9d0a-ec16ead4f421" (UID: "6882541d-9663-41db-9d0a-ec16ead4f421"). InnerVolumeSpecName "kube-api-access-4vcb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.342216 4777 generic.go:334] "Generic (PLEG): container finished" podID="6882541d-9663-41db-9d0a-ec16ead4f421" containerID="c6aed0abf3b15443476bb3d78d14298f34006b46dd8b8b47d5549d7272fb324c" exitCode=0 Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.342294 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6882541d-9663-41db-9d0a-ec16ead4f421","Type":"ContainerDied","Data":"c6aed0abf3b15443476bb3d78d14298f34006b46dd8b8b47d5549d7272fb324c"} Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.342326 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6882541d-9663-41db-9d0a-ec16ead4f421","Type":"ContainerDied","Data":"591832c702a388c65a27f77490095502a9e5762742f2f7fb5d4c24effcc112da"} Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.342347 4777 scope.go:117] "RemoveContainer" containerID="6f8ee2273c0cc1a79dea4f80589dfbbdb693bba486bfd47d18b0c39f0810f381" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.342534 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.347141 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6882541d-9663-41db-9d0a-ec16ead4f421" (UID: "6882541d-9663-41db-9d0a-ec16ead4f421"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.350667 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.350958 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54dd998c-8smx5" event={"ID":"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4","Type":"ContainerStarted","Data":"329084f1a1bf025edf20124bb61ab4512356cda5cfe18ffc22697058d48ceff6"} Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.351119 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.371857 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.371904 4777 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.371920 4777 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6882541d-9663-41db-9d0a-ec16ead4f421-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.371931 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vcb4\" (UniqueName: \"kubernetes.io/projected/6882541d-9663-41db-9d0a-ec16ead4f421-kube-api-access-4vcb4\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.383167 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54dd998c-8smx5" podStartSLOduration=3.3831440329999998 podStartE2EDuration="3.383144033s" podCreationTimestamp="2025-11-24 17:23:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:23:51.370418475 +0000 UTC m=+1389.529613524" watchObservedRunningTime="2025-11-24 17:23:51.383144033 +0000 UTC m=+1389.542339082" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.384796 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6882541d-9663-41db-9d0a-ec16ead4f421" (UID: "6882541d-9663-41db-9d0a-ec16ead4f421"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.423788 4777 scope.go:117] "RemoveContainer" containerID="29d2470b7cadd938824606893257bf5cda27a59d52c06dea02db705e7e6d80ba" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.443433 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.443721 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-config-data" (OuterVolumeSpecName: "config-data") pod "6882541d-9663-41db-9d0a-ec16ead4f421" (UID: "6882541d-9663-41db-9d0a-ec16ead4f421"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.460026 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.469180 4777 scope.go:117] "RemoveContainer" containerID="c6aed0abf3b15443476bb3d78d14298f34006b46dd8b8b47d5549d7272fb324c" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.471549 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:23:51 crc kubenswrapper[4777]: E1124 17:23:51.472031 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67ecd2f1-6485-4d94-b66c-0bc0465ea695" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.472061 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="67ecd2f1-6485-4d94-b66c-0bc0465ea695" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 17:23:51 crc kubenswrapper[4777]: E1124 17:23:51.472075 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="sg-core" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.472082 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="sg-core" Nov 24 17:23:51 crc kubenswrapper[4777]: E1124 17:23:51.472106 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="ceilometer-notification-agent" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.472113 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="ceilometer-notification-agent" Nov 24 17:23:51 crc kubenswrapper[4777]: E1124 17:23:51.472140 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="proxy-httpd" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.472145 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="proxy-httpd" Nov 24 17:23:51 crc kubenswrapper[4777]: E1124 17:23:51.472156 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="ceilometer-central-agent" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.472163 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="ceilometer-central-agent" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.472388 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="67ecd2f1-6485-4d94-b66c-0bc0465ea695" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.472401 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="sg-core" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.472416 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="ceilometer-notification-agent" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.472429 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="proxy-httpd" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.472446 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" containerName="ceilometer-central-agent" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.475126 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.475165 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6882541d-9663-41db-9d0a-ec16ead4f421-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.477440 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.480757 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.483822 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.485006 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.486175 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.507478 4777 scope.go:117] "RemoveContainer" containerID="6e14cb85635d3e244139a9e247c2bce46931d36cabc3d9bd13250d204420dda6" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.526675 4777 scope.go:117] "RemoveContainer" containerID="6f8ee2273c0cc1a79dea4f80589dfbbdb693bba486bfd47d18b0c39f0810f381" Nov 24 17:23:51 crc kubenswrapper[4777]: E1124 17:23:51.527172 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f8ee2273c0cc1a79dea4f80589dfbbdb693bba486bfd47d18b0c39f0810f381\": container with ID starting with 6f8ee2273c0cc1a79dea4f80589dfbbdb693bba486bfd47d18b0c39f0810f381 not found: ID does not exist" containerID="6f8ee2273c0cc1a79dea4f80589dfbbdb693bba486bfd47d18b0c39f0810f381" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.527215 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f8ee2273c0cc1a79dea4f80589dfbbdb693bba486bfd47d18b0c39f0810f381"} err="failed to get container status \"6f8ee2273c0cc1a79dea4f80589dfbbdb693bba486bfd47d18b0c39f0810f381\": rpc error: code = NotFound desc = could not find container \"6f8ee2273c0cc1a79dea4f80589dfbbdb693bba486bfd47d18b0c39f0810f381\": container with ID starting with 6f8ee2273c0cc1a79dea4f80589dfbbdb693bba486bfd47d18b0c39f0810f381 not found: ID does not exist" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.527250 4777 scope.go:117] "RemoveContainer" containerID="29d2470b7cadd938824606893257bf5cda27a59d52c06dea02db705e7e6d80ba" Nov 24 17:23:51 crc kubenswrapper[4777]: E1124 17:23:51.527617 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29d2470b7cadd938824606893257bf5cda27a59d52c06dea02db705e7e6d80ba\": container with ID starting with 29d2470b7cadd938824606893257bf5cda27a59d52c06dea02db705e7e6d80ba not found: ID does not exist" containerID="29d2470b7cadd938824606893257bf5cda27a59d52c06dea02db705e7e6d80ba" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.527648 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29d2470b7cadd938824606893257bf5cda27a59d52c06dea02db705e7e6d80ba"} err="failed to get container status \"29d2470b7cadd938824606893257bf5cda27a59d52c06dea02db705e7e6d80ba\": rpc error: code = NotFound desc = could not find container \"29d2470b7cadd938824606893257bf5cda27a59d52c06dea02db705e7e6d80ba\": container with ID starting with 29d2470b7cadd938824606893257bf5cda27a59d52c06dea02db705e7e6d80ba not found: ID does not exist" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.527669 4777 scope.go:117] "RemoveContainer" containerID="c6aed0abf3b15443476bb3d78d14298f34006b46dd8b8b47d5549d7272fb324c" Nov 24 17:23:51 crc kubenswrapper[4777]: E1124 17:23:51.528181 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6aed0abf3b15443476bb3d78d14298f34006b46dd8b8b47d5549d7272fb324c\": container with ID starting with c6aed0abf3b15443476bb3d78d14298f34006b46dd8b8b47d5549d7272fb324c not found: ID does not exist" containerID="c6aed0abf3b15443476bb3d78d14298f34006b46dd8b8b47d5549d7272fb324c" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.528201 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6aed0abf3b15443476bb3d78d14298f34006b46dd8b8b47d5549d7272fb324c"} err="failed to get container status \"c6aed0abf3b15443476bb3d78d14298f34006b46dd8b8b47d5549d7272fb324c\": rpc error: code = NotFound desc = could not find container \"c6aed0abf3b15443476bb3d78d14298f34006b46dd8b8b47d5549d7272fb324c\": container with ID starting with c6aed0abf3b15443476bb3d78d14298f34006b46dd8b8b47d5549d7272fb324c not found: ID does not exist" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.528216 4777 scope.go:117] "RemoveContainer" containerID="6e14cb85635d3e244139a9e247c2bce46931d36cabc3d9bd13250d204420dda6" Nov 24 17:23:51 crc kubenswrapper[4777]: E1124 17:23:51.528485 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e14cb85635d3e244139a9e247c2bce46931d36cabc3d9bd13250d204420dda6\": container with ID starting with 6e14cb85635d3e244139a9e247c2bce46931d36cabc3d9bd13250d204420dda6 not found: ID does not exist" containerID="6e14cb85635d3e244139a9e247c2bce46931d36cabc3d9bd13250d204420dda6" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.528515 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e14cb85635d3e244139a9e247c2bce46931d36cabc3d9bd13250d204420dda6"} err="failed to get container status \"6e14cb85635d3e244139a9e247c2bce46931d36cabc3d9bd13250d204420dda6\": rpc error: code = NotFound desc = could not find container \"6e14cb85635d3e244139a9e247c2bce46931d36cabc3d9bd13250d204420dda6\": container with ID starting with 6e14cb85635d3e244139a9e247c2bce46931d36cabc3d9bd13250d204420dda6 not found: ID does not exist" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.577010 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ctgm\" (UniqueName: \"kubernetes.io/projected/ad8c715b-ea28-4366-882c-4371692973f0-kube-api-access-2ctgm\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad8c715b-ea28-4366-882c-4371692973f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.577106 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad8c715b-ea28-4366-882c-4371692973f0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad8c715b-ea28-4366-882c-4371692973f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.577155 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad8c715b-ea28-4366-882c-4371692973f0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad8c715b-ea28-4366-882c-4371692973f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.577187 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad8c715b-ea28-4366-882c-4371692973f0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad8c715b-ea28-4366-882c-4371692973f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.577257 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad8c715b-ea28-4366-882c-4371692973f0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad8c715b-ea28-4366-882c-4371692973f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.674245 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.679704 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad8c715b-ea28-4366-882c-4371692973f0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad8c715b-ea28-4366-882c-4371692973f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.679943 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ctgm\" (UniqueName: \"kubernetes.io/projected/ad8c715b-ea28-4366-882c-4371692973f0-kube-api-access-2ctgm\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad8c715b-ea28-4366-882c-4371692973f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.680005 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad8c715b-ea28-4366-882c-4371692973f0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad8c715b-ea28-4366-882c-4371692973f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.680050 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad8c715b-ea28-4366-882c-4371692973f0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad8c715b-ea28-4366-882c-4371692973f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.680081 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad8c715b-ea28-4366-882c-4371692973f0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad8c715b-ea28-4366-882c-4371692973f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.683696 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad8c715b-ea28-4366-882c-4371692973f0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad8c715b-ea28-4366-882c-4371692973f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.684121 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad8c715b-ea28-4366-882c-4371692973f0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad8c715b-ea28-4366-882c-4371692973f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.684526 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad8c715b-ea28-4366-882c-4371692973f0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad8c715b-ea28-4366-882c-4371692973f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.685028 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad8c715b-ea28-4366-882c-4371692973f0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad8c715b-ea28-4366-882c-4371692973f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.685864 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.700851 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ctgm\" (UniqueName: \"kubernetes.io/projected/ad8c715b-ea28-4366-882c-4371692973f0-kube-api-access-2ctgm\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad8c715b-ea28-4366-882c-4371692973f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.709106 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.711570 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.713816 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.714130 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.714261 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.722876 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.785848 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-config-data\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.786502 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.786611 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-run-httpd\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.786730 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-scripts\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.786932 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.787161 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p9w8\" (UniqueName: \"kubernetes.io/projected/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-kube-api-access-9p9w8\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.787247 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.787311 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-log-httpd\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.803220 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.820193 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.820556 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="69de6629-9de1-41c1-aa41-4bb079a0505c" containerName="nova-api-log" containerID="cri-o://86b79550368087fc8250f42d992d4f72846badbbc81795a5d351e4c7fb82b53d" gracePeriod=30 Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.821304 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="69de6629-9de1-41c1-aa41-4bb079a0505c" containerName="nova-api-api" containerID="cri-o://fc4e8352af11ae0b15300cb0dc91ff08b6acc39f8ebf5948877c6b153e61dbc7" gracePeriod=30 Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.894107 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-log-httpd\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.894551 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-config-data\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.894622 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.894645 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-run-httpd\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.894702 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-scripts\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.894735 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.894799 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p9w8\" (UniqueName: \"kubernetes.io/projected/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-kube-api-access-9p9w8\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.894839 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.895776 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-log-httpd\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.895814 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-run-httpd\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.899378 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.901959 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-scripts\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.906555 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-config-data\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.906636 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.908712 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:51 crc kubenswrapper[4777]: I1124 17:23:51.923303 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p9w8\" (UniqueName: \"kubernetes.io/projected/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-kube-api-access-9p9w8\") pod \"ceilometer-0\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " pod="openstack/ceilometer-0" Nov 24 17:23:52 crc kubenswrapper[4777]: I1124 17:23:52.046647 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:23:52 crc kubenswrapper[4777]: I1124 17:23:52.364265 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ad8c715b-ea28-4366-882c-4371692973f0","Type":"ContainerStarted","Data":"40749e07c5775ce68aa190a02341cc9a0b4064405d2c0d136190a3a82428b276"} Nov 24 17:23:52 crc kubenswrapper[4777]: I1124 17:23:52.370478 4777 generic.go:334] "Generic (PLEG): container finished" podID="69de6629-9de1-41c1-aa41-4bb079a0505c" containerID="86b79550368087fc8250f42d992d4f72846badbbc81795a5d351e4c7fb82b53d" exitCode=143 Nov 24 17:23:52 crc kubenswrapper[4777]: I1124 17:23:52.371385 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"69de6629-9de1-41c1-aa41-4bb079a0505c","Type":"ContainerDied","Data":"86b79550368087fc8250f42d992d4f72846badbbc81795a5d351e4c7fb82b53d"} Nov 24 17:23:52 crc kubenswrapper[4777]: I1124 17:23:52.371420 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 17:23:52 crc kubenswrapper[4777]: I1124 17:23:52.389919 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:23:52 crc kubenswrapper[4777]: I1124 17:23:52.534246 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:23:52 crc kubenswrapper[4777]: W1124 17:23:52.537211 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56cb23b8_22f8_4e1a_9db5_7d75ee69114d.slice/crio-bce3bb08725129d9a04e8d01bbba4cc0502b0f7abbeabad359672c8599017c35 WatchSource:0}: Error finding container bce3bb08725129d9a04e8d01bbba4cc0502b0f7abbeabad359672c8599017c35: Status 404 returned error can't find the container with id bce3bb08725129d9a04e8d01bbba4cc0502b0f7abbeabad359672c8599017c35 Nov 24 17:23:53 crc kubenswrapper[4777]: I1124 17:23:53.265242 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67ecd2f1-6485-4d94-b66c-0bc0465ea695" path="/var/lib/kubelet/pods/67ecd2f1-6485-4d94-b66c-0bc0465ea695/volumes" Nov 24 17:23:53 crc kubenswrapper[4777]: I1124 17:23:53.266502 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6882541d-9663-41db-9d0a-ec16ead4f421" path="/var/lib/kubelet/pods/6882541d-9663-41db-9d0a-ec16ead4f421/volumes" Nov 24 17:23:53 crc kubenswrapper[4777]: I1124 17:23:53.382219 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ad8c715b-ea28-4366-882c-4371692973f0","Type":"ContainerStarted","Data":"6684199f66e05b22f4fca87c96650878fc8139bc480a5802da979269045b6ce5"} Nov 24 17:23:53 crc kubenswrapper[4777]: I1124 17:23:53.383817 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56cb23b8-22f8-4e1a-9db5-7d75ee69114d","Type":"ContainerStarted","Data":"347879372d7574fe36c49df16be9ff348e75e7e85ac44c200db592e69d90fe1d"} Nov 24 17:23:53 crc kubenswrapper[4777]: I1124 17:23:53.383853 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56cb23b8-22f8-4e1a-9db5-7d75ee69114d","Type":"ContainerStarted","Data":"bce3bb08725129d9a04e8d01bbba4cc0502b0f7abbeabad359672c8599017c35"} Nov 24 17:23:53 crc kubenswrapper[4777]: I1124 17:23:53.406884 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.4068542649999998 podStartE2EDuration="2.406854265s" podCreationTimestamp="2025-11-24 17:23:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:23:53.397870195 +0000 UTC m=+1391.557065244" watchObservedRunningTime="2025-11-24 17:23:53.406854265 +0000 UTC m=+1391.566049334" Nov 24 17:23:54 crc kubenswrapper[4777]: I1124 17:23:54.396660 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56cb23b8-22f8-4e1a-9db5-7d75ee69114d","Type":"ContainerStarted","Data":"5f44b2189bacf51f378c9bb20f6636a8102ea448f2d8cbabef077dd9082e9109"} Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.436173 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56cb23b8-22f8-4e1a-9db5-7d75ee69114d","Type":"ContainerStarted","Data":"6eec3125eb7adf1454d8a4f71572df52f99cbdd88a321a71f16bc3ad7aff8d1a"} Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.438364 4777 generic.go:334] "Generic (PLEG): container finished" podID="69de6629-9de1-41c1-aa41-4bb079a0505c" containerID="fc4e8352af11ae0b15300cb0dc91ff08b6acc39f8ebf5948877c6b153e61dbc7" exitCode=0 Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.438414 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"69de6629-9de1-41c1-aa41-4bb079a0505c","Type":"ContainerDied","Data":"fc4e8352af11ae0b15300cb0dc91ff08b6acc39f8ebf5948877c6b153e61dbc7"} Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.588522 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.681106 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69de6629-9de1-41c1-aa41-4bb079a0505c-config-data\") pod \"69de6629-9de1-41c1-aa41-4bb079a0505c\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.681299 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdl8x\" (UniqueName: \"kubernetes.io/projected/69de6629-9de1-41c1-aa41-4bb079a0505c-kube-api-access-tdl8x\") pod \"69de6629-9de1-41c1-aa41-4bb079a0505c\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.681343 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69de6629-9de1-41c1-aa41-4bb079a0505c-combined-ca-bundle\") pod \"69de6629-9de1-41c1-aa41-4bb079a0505c\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.681471 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69de6629-9de1-41c1-aa41-4bb079a0505c-logs\") pod \"69de6629-9de1-41c1-aa41-4bb079a0505c\" (UID: \"69de6629-9de1-41c1-aa41-4bb079a0505c\") " Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.683930 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69de6629-9de1-41c1-aa41-4bb079a0505c-logs" (OuterVolumeSpecName: "logs") pod "69de6629-9de1-41c1-aa41-4bb079a0505c" (UID: "69de6629-9de1-41c1-aa41-4bb079a0505c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.691185 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69de6629-9de1-41c1-aa41-4bb079a0505c-kube-api-access-tdl8x" (OuterVolumeSpecName: "kube-api-access-tdl8x") pod "69de6629-9de1-41c1-aa41-4bb079a0505c" (UID: "69de6629-9de1-41c1-aa41-4bb079a0505c"). InnerVolumeSpecName "kube-api-access-tdl8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.708151 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.732026 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69de6629-9de1-41c1-aa41-4bb079a0505c-config-data" (OuterVolumeSpecName: "config-data") pod "69de6629-9de1-41c1-aa41-4bb079a0505c" (UID: "69de6629-9de1-41c1-aa41-4bb079a0505c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.749120 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69de6629-9de1-41c1-aa41-4bb079a0505c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69de6629-9de1-41c1-aa41-4bb079a0505c" (UID: "69de6629-9de1-41c1-aa41-4bb079a0505c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.783578 4777 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69de6629-9de1-41c1-aa41-4bb079a0505c-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.783920 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69de6629-9de1-41c1-aa41-4bb079a0505c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.783936 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdl8x\" (UniqueName: \"kubernetes.io/projected/69de6629-9de1-41c1-aa41-4bb079a0505c-kube-api-access-tdl8x\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:55 crc kubenswrapper[4777]: I1124 17:23:55.783952 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69de6629-9de1-41c1-aa41-4bb079a0505c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.461526 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"69de6629-9de1-41c1-aa41-4bb079a0505c","Type":"ContainerDied","Data":"2cce1bc7fd0c41bc72f125a2975c1eda218a1a79194f68b899aae81aa979036f"} Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.461896 4777 scope.go:117] "RemoveContainer" containerID="fc4e8352af11ae0b15300cb0dc91ff08b6acc39f8ebf5948877c6b153e61dbc7" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.461622 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.492402 4777 scope.go:117] "RemoveContainer" containerID="86b79550368087fc8250f42d992d4f72846badbbc81795a5d351e4c7fb82b53d" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.496558 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.505577 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.523764 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 17:23:56 crc kubenswrapper[4777]: E1124 17:23:56.524203 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69de6629-9de1-41c1-aa41-4bb079a0505c" containerName="nova-api-api" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.524219 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="69de6629-9de1-41c1-aa41-4bb079a0505c" containerName="nova-api-api" Nov 24 17:23:56 crc kubenswrapper[4777]: E1124 17:23:56.524234 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69de6629-9de1-41c1-aa41-4bb079a0505c" containerName="nova-api-log" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.524322 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="69de6629-9de1-41c1-aa41-4bb079a0505c" containerName="nova-api-log" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.524580 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="69de6629-9de1-41c1-aa41-4bb079a0505c" containerName="nova-api-log" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.524617 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="69de6629-9de1-41c1-aa41-4bb079a0505c" containerName="nova-api-api" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.525741 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.529692 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.529983 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.530153 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.551171 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.704841 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-logs\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.704934 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.705035 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-public-tls-certs\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.705089 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-config-data\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.705157 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd2bc\" (UniqueName: \"kubernetes.io/projected/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-kube-api-access-cd2bc\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.705242 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.805078 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.806646 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.806727 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-public-tls-certs\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.806784 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-config-data\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.806834 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd2bc\" (UniqueName: \"kubernetes.io/projected/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-kube-api-access-cd2bc\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.806923 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.806997 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-logs\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.807697 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-logs\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.811751 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.812228 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-public-tls-certs\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.813576 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-config-data\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.826781 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.832277 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd2bc\" (UniqueName: \"kubernetes.io/projected/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-kube-api-access-cd2bc\") pod \"nova-api-0\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " pod="openstack/nova-api-0" Nov 24 17:23:56 crc kubenswrapper[4777]: I1124 17:23:56.854301 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:23:57 crc kubenswrapper[4777]: I1124 17:23:57.262187 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69de6629-9de1-41c1-aa41-4bb079a0505c" path="/var/lib/kubelet/pods/69de6629-9de1-41c1-aa41-4bb079a0505c/volumes" Nov 24 17:23:57 crc kubenswrapper[4777]: I1124 17:23:57.387651 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:23:57 crc kubenswrapper[4777]: I1124 17:23:57.477370 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56cb23b8-22f8-4e1a-9db5-7d75ee69114d","Type":"ContainerStarted","Data":"c2f888860eec03536a3258c6cd3b4bb66f2d65b8a679cacc23fc478e75ba677b"} Nov 24 17:23:57 crc kubenswrapper[4777]: I1124 17:23:57.477538 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="ceilometer-central-agent" containerID="cri-o://347879372d7574fe36c49df16be9ff348e75e7e85ac44c200db592e69d90fe1d" gracePeriod=30 Nov 24 17:23:57 crc kubenswrapper[4777]: I1124 17:23:57.477805 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 17:23:57 crc kubenswrapper[4777]: I1124 17:23:57.477803 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="proxy-httpd" containerID="cri-o://c2f888860eec03536a3258c6cd3b4bb66f2d65b8a679cacc23fc478e75ba677b" gracePeriod=30 Nov 24 17:23:57 crc kubenswrapper[4777]: I1124 17:23:57.477822 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="sg-core" containerID="cri-o://6eec3125eb7adf1454d8a4f71572df52f99cbdd88a321a71f16bc3ad7aff8d1a" gracePeriod=30 Nov 24 17:23:57 crc kubenswrapper[4777]: I1124 17:23:57.477872 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="ceilometer-notification-agent" containerID="cri-o://5f44b2189bacf51f378c9bb20f6636a8102ea448f2d8cbabef077dd9082e9109" gracePeriod=30 Nov 24 17:23:57 crc kubenswrapper[4777]: I1124 17:23:57.493335 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e","Type":"ContainerStarted","Data":"af9cf68ef7847f2dc8e00ea29beb884011222ba9a53c420406cfbaeca88be471"} Nov 24 17:23:58 crc kubenswrapper[4777]: I1124 17:23:58.506097 4777 generic.go:334] "Generic (PLEG): container finished" podID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerID="c2f888860eec03536a3258c6cd3b4bb66f2d65b8a679cacc23fc478e75ba677b" exitCode=0 Nov 24 17:23:58 crc kubenswrapper[4777]: I1124 17:23:58.506674 4777 generic.go:334] "Generic (PLEG): container finished" podID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerID="6eec3125eb7adf1454d8a4f71572df52f99cbdd88a321a71f16bc3ad7aff8d1a" exitCode=2 Nov 24 17:23:58 crc kubenswrapper[4777]: I1124 17:23:58.506688 4777 generic.go:334] "Generic (PLEG): container finished" podID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerID="5f44b2189bacf51f378c9bb20f6636a8102ea448f2d8cbabef077dd9082e9109" exitCode=0 Nov 24 17:23:58 crc kubenswrapper[4777]: I1124 17:23:58.506172 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56cb23b8-22f8-4e1a-9db5-7d75ee69114d","Type":"ContainerDied","Data":"c2f888860eec03536a3258c6cd3b4bb66f2d65b8a679cacc23fc478e75ba677b"} Nov 24 17:23:58 crc kubenswrapper[4777]: I1124 17:23:58.506820 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56cb23b8-22f8-4e1a-9db5-7d75ee69114d","Type":"ContainerDied","Data":"6eec3125eb7adf1454d8a4f71572df52f99cbdd88a321a71f16bc3ad7aff8d1a"} Nov 24 17:23:58 crc kubenswrapper[4777]: I1124 17:23:58.506854 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56cb23b8-22f8-4e1a-9db5-7d75ee69114d","Type":"ContainerDied","Data":"5f44b2189bacf51f378c9bb20f6636a8102ea448f2d8cbabef077dd9082e9109"} Nov 24 17:23:58 crc kubenswrapper[4777]: I1124 17:23:58.509472 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e","Type":"ContainerStarted","Data":"780deec4df91e72ee6fb8e88db0289a8bce65141a13ec2700f0c77e5d58a6068"} Nov 24 17:23:58 crc kubenswrapper[4777]: I1124 17:23:58.509498 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e","Type":"ContainerStarted","Data":"a0cd6a9f01d4e328ed595112a30fd680948d429e10c36688f79eddaaf8c213b6"} Nov 24 17:23:58 crc kubenswrapper[4777]: I1124 17:23:58.539444 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.802704159 podStartE2EDuration="7.539426128s" podCreationTimestamp="2025-11-24 17:23:51 +0000 UTC" firstStartedPulling="2025-11-24 17:23:52.53942307 +0000 UTC m=+1390.698618119" lastFinishedPulling="2025-11-24 17:23:56.276145029 +0000 UTC m=+1394.435340088" observedRunningTime="2025-11-24 17:23:57.498034854 +0000 UTC m=+1395.657229913" watchObservedRunningTime="2025-11-24 17:23:58.539426128 +0000 UTC m=+1396.698621177" Nov 24 17:23:58 crc kubenswrapper[4777]: I1124 17:23:58.541418 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.541410245 podStartE2EDuration="2.541410245s" podCreationTimestamp="2025-11-24 17:23:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:23:58.534802254 +0000 UTC m=+1396.693997303" watchObservedRunningTime="2025-11-24 17:23:58.541410245 +0000 UTC m=+1396.700605294" Nov 24 17:23:58 crc kubenswrapper[4777]: I1124 17:23:58.838119 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:23:58 crc kubenswrapper[4777]: I1124 17:23:58.939097 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-884c8b8f5-v85nr"] Nov 24 17:23:58 crc kubenswrapper[4777]: I1124 17:23:58.939407 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" podUID="8a758074-c834-4b50-8f0e-8959700e9cab" containerName="dnsmasq-dns" containerID="cri-o://287001c4ae3dcc885df3c7ccad15a8be02ef267473fdfc38ee67728123ab4169" gracePeriod=10 Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.526422 4777 generic.go:334] "Generic (PLEG): container finished" podID="8a758074-c834-4b50-8f0e-8959700e9cab" containerID="287001c4ae3dcc885df3c7ccad15a8be02ef267473fdfc38ee67728123ab4169" exitCode=0 Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.526489 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" event={"ID":"8a758074-c834-4b50-8f0e-8959700e9cab","Type":"ContainerDied","Data":"287001c4ae3dcc885df3c7ccad15a8be02ef267473fdfc38ee67728123ab4169"} Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.681111 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.787289 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-dns-svc\") pod \"8a758074-c834-4b50-8f0e-8959700e9cab\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.787454 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-ovsdbserver-sb\") pod \"8a758074-c834-4b50-8f0e-8959700e9cab\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.787501 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-ovsdbserver-nb\") pod \"8a758074-c834-4b50-8f0e-8959700e9cab\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.787525 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcp7t\" (UniqueName: \"kubernetes.io/projected/8a758074-c834-4b50-8f0e-8959700e9cab-kube-api-access-dcp7t\") pod \"8a758074-c834-4b50-8f0e-8959700e9cab\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.787585 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-dns-swift-storage-0\") pod \"8a758074-c834-4b50-8f0e-8959700e9cab\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.787750 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-config\") pod \"8a758074-c834-4b50-8f0e-8959700e9cab\" (UID: \"8a758074-c834-4b50-8f0e-8959700e9cab\") " Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.807211 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a758074-c834-4b50-8f0e-8959700e9cab-kube-api-access-dcp7t" (OuterVolumeSpecName: "kube-api-access-dcp7t") pod "8a758074-c834-4b50-8f0e-8959700e9cab" (UID: "8a758074-c834-4b50-8f0e-8959700e9cab"). InnerVolumeSpecName "kube-api-access-dcp7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.853657 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-config" (OuterVolumeSpecName: "config") pod "8a758074-c834-4b50-8f0e-8959700e9cab" (UID: "8a758074-c834-4b50-8f0e-8959700e9cab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.864674 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8a758074-c834-4b50-8f0e-8959700e9cab" (UID: "8a758074-c834-4b50-8f0e-8959700e9cab"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.869559 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8a758074-c834-4b50-8f0e-8959700e9cab" (UID: "8a758074-c834-4b50-8f0e-8959700e9cab"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.871258 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8a758074-c834-4b50-8f0e-8959700e9cab" (UID: "8a758074-c834-4b50-8f0e-8959700e9cab"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.878924 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8a758074-c834-4b50-8f0e-8959700e9cab" (UID: "8a758074-c834-4b50-8f0e-8959700e9cab"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.891832 4777 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.892048 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.892113 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.892172 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.892264 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a758074-c834-4b50-8f0e-8959700e9cab-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:23:59 crc kubenswrapper[4777]: I1124 17:23:59.892321 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcp7t\" (UniqueName: \"kubernetes.io/projected/8a758074-c834-4b50-8f0e-8959700e9cab-kube-api-access-dcp7t\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:00 crc kubenswrapper[4777]: I1124 17:24:00.539108 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" event={"ID":"8a758074-c834-4b50-8f0e-8959700e9cab","Type":"ContainerDied","Data":"d936cddb2620ace12eee0a90c826a590079b5e1eebabc84ceb41cbb9985ed9ef"} Nov 24 17:24:00 crc kubenswrapper[4777]: I1124 17:24:00.539573 4777 scope.go:117] "RemoveContainer" containerID="287001c4ae3dcc885df3c7ccad15a8be02ef267473fdfc38ee67728123ab4169" Nov 24 17:24:00 crc kubenswrapper[4777]: I1124 17:24:00.539240 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-884c8b8f5-v85nr" Nov 24 17:24:00 crc kubenswrapper[4777]: I1124 17:24:00.579566 4777 scope.go:117] "RemoveContainer" containerID="6747216930bd90bd130aea78411607661ebcfdfe42532a16dfb7485da87dd047" Nov 24 17:24:00 crc kubenswrapper[4777]: I1124 17:24:00.580786 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-884c8b8f5-v85nr"] Nov 24 17:24:00 crc kubenswrapper[4777]: I1124 17:24:00.591247 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-884c8b8f5-v85nr"] Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.258764 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a758074-c834-4b50-8f0e-8959700e9cab" path="/var/lib/kubelet/pods/8a758074-c834-4b50-8f0e-8959700e9cab/volumes" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.551665 4777 generic.go:334] "Generic (PLEG): container finished" podID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerID="347879372d7574fe36c49df16be9ff348e75e7e85ac44c200db592e69d90fe1d" exitCode=0 Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.551698 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56cb23b8-22f8-4e1a-9db5-7d75ee69114d","Type":"ContainerDied","Data":"347879372d7574fe36c49df16be9ff348e75e7e85ac44c200db592e69d90fe1d"} Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.674739 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.805112 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.826799 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.828052 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-log-httpd\") pod \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.828131 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-config-data\") pod \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.828276 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-ceilometer-tls-certs\") pod \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.828443 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-scripts\") pod \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.828477 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-combined-ca-bundle\") pod \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.828556 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-run-httpd\") pod \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.828628 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9p9w8\" (UniqueName: \"kubernetes.io/projected/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-kube-api-access-9p9w8\") pod \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.828668 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-sg-core-conf-yaml\") pod \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\" (UID: \"56cb23b8-22f8-4e1a-9db5-7d75ee69114d\") " Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.828478 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "56cb23b8-22f8-4e1a-9db5-7d75ee69114d" (UID: "56cb23b8-22f8-4e1a-9db5-7d75ee69114d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.829287 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "56cb23b8-22f8-4e1a-9db5-7d75ee69114d" (UID: "56cb23b8-22f8-4e1a-9db5-7d75ee69114d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.829949 4777 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.830003 4777 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.834741 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-kube-api-access-9p9w8" (OuterVolumeSpecName: "kube-api-access-9p9w8") pod "56cb23b8-22f8-4e1a-9db5-7d75ee69114d" (UID: "56cb23b8-22f8-4e1a-9db5-7d75ee69114d"). InnerVolumeSpecName "kube-api-access-9p9w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.835564 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-scripts" (OuterVolumeSpecName: "scripts") pod "56cb23b8-22f8-4e1a-9db5-7d75ee69114d" (UID: "56cb23b8-22f8-4e1a-9db5-7d75ee69114d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.877163 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "56cb23b8-22f8-4e1a-9db5-7d75ee69114d" (UID: "56cb23b8-22f8-4e1a-9db5-7d75ee69114d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.931842 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.931876 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9p9w8\" (UniqueName: \"kubernetes.io/projected/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-kube-api-access-9p9w8\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.931890 4777 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.934371 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "56cb23b8-22f8-4e1a-9db5-7d75ee69114d" (UID: "56cb23b8-22f8-4e1a-9db5-7d75ee69114d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.941153 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56cb23b8-22f8-4e1a-9db5-7d75ee69114d" (UID: "56cb23b8-22f8-4e1a-9db5-7d75ee69114d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:01 crc kubenswrapper[4777]: I1124 17:24:01.979166 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-config-data" (OuterVolumeSpecName: "config-data") pod "56cb23b8-22f8-4e1a-9db5-7d75ee69114d" (UID: "56cb23b8-22f8-4e1a-9db5-7d75ee69114d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.033760 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.033797 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.033805 4777 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/56cb23b8-22f8-4e1a-9db5-7d75ee69114d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.566575 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.568115 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56cb23b8-22f8-4e1a-9db5-7d75ee69114d","Type":"ContainerDied","Data":"bce3bb08725129d9a04e8d01bbba4cc0502b0f7abbeabad359672c8599017c35"} Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.568170 4777 scope.go:117] "RemoveContainer" containerID="c2f888860eec03536a3258c6cd3b4bb66f2d65b8a679cacc23fc478e75ba677b" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.585932 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.614260 4777 scope.go:117] "RemoveContainer" containerID="6eec3125eb7adf1454d8a4f71572df52f99cbdd88a321a71f16bc3ad7aff8d1a" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.642069 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.657518 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.671728 4777 scope.go:117] "RemoveContainer" containerID="5f44b2189bacf51f378c9bb20f6636a8102ea448f2d8cbabef077dd9082e9109" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.671862 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:24:02 crc kubenswrapper[4777]: E1124 17:24:02.672292 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a758074-c834-4b50-8f0e-8959700e9cab" containerName="init" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.672308 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a758074-c834-4b50-8f0e-8959700e9cab" containerName="init" Nov 24 17:24:02 crc kubenswrapper[4777]: E1124 17:24:02.672317 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a758074-c834-4b50-8f0e-8959700e9cab" containerName="dnsmasq-dns" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.672323 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a758074-c834-4b50-8f0e-8959700e9cab" containerName="dnsmasq-dns" Nov 24 17:24:02 crc kubenswrapper[4777]: E1124 17:24:02.672342 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="sg-core" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.672348 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="sg-core" Nov 24 17:24:02 crc kubenswrapper[4777]: E1124 17:24:02.672357 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="ceilometer-central-agent" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.672362 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="ceilometer-central-agent" Nov 24 17:24:02 crc kubenswrapper[4777]: E1124 17:24:02.672372 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="ceilometer-notification-agent" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.672378 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="ceilometer-notification-agent" Nov 24 17:24:02 crc kubenswrapper[4777]: E1124 17:24:02.672390 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="proxy-httpd" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.672395 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="proxy-httpd" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.672599 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="ceilometer-central-agent" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.672619 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a758074-c834-4b50-8f0e-8959700e9cab" containerName="dnsmasq-dns" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.672626 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="proxy-httpd" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.672643 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="ceilometer-notification-agent" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.672650 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" containerName="sg-core" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.674470 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.686570 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.686794 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.686895 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.692136 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.723847 4777 scope.go:117] "RemoveContainer" containerID="347879372d7574fe36c49df16be9ff348e75e7e85ac44c200db592e69d90fe1d" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.847931 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-rjw7h"] Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.849585 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.850778 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.850819 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9412f05c-2d23-4124-abde-2f573bb6855d-run-httpd\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.850847 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-config-data\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.851141 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.851372 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9412f05c-2d23-4124-abde-2f573bb6855d-log-httpd\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.851405 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.851445 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-scripts\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.851581 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnd4b\" (UniqueName: \"kubernetes.io/projected/9412f05c-2d23-4124-abde-2f573bb6855d-kube-api-access-xnd4b\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.851756 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.859852 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.861824 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-rjw7h"] Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.953110 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.953174 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9412f05c-2d23-4124-abde-2f573bb6855d-log-httpd\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.953204 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.953228 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-scripts\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.953316 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnd4b\" (UniqueName: \"kubernetes.io/projected/9412f05c-2d23-4124-abde-2f573bb6855d-kube-api-access-xnd4b\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.953349 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rjw7h\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.953914 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-config-data\") pod \"nova-cell1-cell-mapping-rjw7h\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.954151 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmvnn\" (UniqueName: \"kubernetes.io/projected/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-kube-api-access-wmvnn\") pod \"nova-cell1-cell-mapping-rjw7h\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.954216 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.954244 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9412f05c-2d23-4124-abde-2f573bb6855d-run-httpd\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.954275 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-config-data\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.954340 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-scripts\") pod \"nova-cell1-cell-mapping-rjw7h\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.954357 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9412f05c-2d23-4124-abde-2f573bb6855d-log-httpd\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.955593 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9412f05c-2d23-4124-abde-2f573bb6855d-run-httpd\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.958942 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-scripts\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.959275 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.960171 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-config-data\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.962325 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.971836 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:02 crc kubenswrapper[4777]: I1124 17:24:02.976132 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnd4b\" (UniqueName: \"kubernetes.io/projected/9412f05c-2d23-4124-abde-2f573bb6855d-kube-api-access-xnd4b\") pod \"ceilometer-0\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " pod="openstack/ceilometer-0" Nov 24 17:24:03 crc kubenswrapper[4777]: I1124 17:24:03.016029 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:24:03 crc kubenswrapper[4777]: I1124 17:24:03.056417 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rjw7h\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:03 crc kubenswrapper[4777]: I1124 17:24:03.056479 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-config-data\") pod \"nova-cell1-cell-mapping-rjw7h\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:03 crc kubenswrapper[4777]: I1124 17:24:03.056560 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmvnn\" (UniqueName: \"kubernetes.io/projected/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-kube-api-access-wmvnn\") pod \"nova-cell1-cell-mapping-rjw7h\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:03 crc kubenswrapper[4777]: I1124 17:24:03.056639 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-scripts\") pod \"nova-cell1-cell-mapping-rjw7h\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:03 crc kubenswrapper[4777]: I1124 17:24:03.061753 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-scripts\") pod \"nova-cell1-cell-mapping-rjw7h\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:03 crc kubenswrapper[4777]: I1124 17:24:03.063322 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rjw7h\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:03 crc kubenswrapper[4777]: I1124 17:24:03.063677 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-config-data\") pod \"nova-cell1-cell-mapping-rjw7h\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:03 crc kubenswrapper[4777]: I1124 17:24:03.080107 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmvnn\" (UniqueName: \"kubernetes.io/projected/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-kube-api-access-wmvnn\") pod \"nova-cell1-cell-mapping-rjw7h\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:03 crc kubenswrapper[4777]: I1124 17:24:03.170436 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:03 crc kubenswrapper[4777]: I1124 17:24:03.285692 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56cb23b8-22f8-4e1a-9db5-7d75ee69114d" path="/var/lib/kubelet/pods/56cb23b8-22f8-4e1a-9db5-7d75ee69114d/volumes" Nov 24 17:24:03 crc kubenswrapper[4777]: I1124 17:24:03.543897 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:24:03 crc kubenswrapper[4777]: I1124 17:24:03.589951 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9412f05c-2d23-4124-abde-2f573bb6855d","Type":"ContainerStarted","Data":"219de3d81339f912e2e280a46dd2e93b9a6f59dc339241204d410f575d9e5330"} Nov 24 17:24:03 crc kubenswrapper[4777]: I1124 17:24:03.684851 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-rjw7h"] Nov 24 17:24:04 crc kubenswrapper[4777]: I1124 17:24:04.611160 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rjw7h" event={"ID":"2c1082e3-ae04-40f7-83ae-f7b7299b26b8","Type":"ContainerStarted","Data":"1b45413f849b6a67ffd7d2107e94d990e4106b4558326e9226a3e9c6a4c03f4d"} Nov 24 17:24:04 crc kubenswrapper[4777]: I1124 17:24:04.611370 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rjw7h" event={"ID":"2c1082e3-ae04-40f7-83ae-f7b7299b26b8","Type":"ContainerStarted","Data":"cd3204032522c06f9d50f81f8e42aaeb8b9a791fb67cda5fc6469cbf2929e135"} Nov 24 17:24:05 crc kubenswrapper[4777]: I1124 17:24:05.623591 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9412f05c-2d23-4124-abde-2f573bb6855d","Type":"ContainerStarted","Data":"f5bc687fc5668ac4935952f9e02557d75c41fa5a98550fa1786c9d926584e7ae"} Nov 24 17:24:06 crc kubenswrapper[4777]: I1124 17:24:06.637877 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9412f05c-2d23-4124-abde-2f573bb6855d","Type":"ContainerStarted","Data":"1a8c50da8aeb8068e5d4150d334c3b86a8ba7c7a17681cd61776dfe674563d98"} Nov 24 17:24:06 crc kubenswrapper[4777]: I1124 17:24:06.638225 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9412f05c-2d23-4124-abde-2f573bb6855d","Type":"ContainerStarted","Data":"117c081d4641d4be6b7400e1d344b0278b56d84856124fa09bc36ccec9a2698e"} Nov 24 17:24:06 crc kubenswrapper[4777]: I1124 17:24:06.854867 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 17:24:06 crc kubenswrapper[4777]: I1124 17:24:06.854911 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 17:24:07 crc kubenswrapper[4777]: I1124 17:24:07.896876 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.225:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 17:24:07 crc kubenswrapper[4777]: I1124 17:24:07.897027 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.225:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 17:24:08 crc kubenswrapper[4777]: I1124 17:24:08.668918 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9412f05c-2d23-4124-abde-2f573bb6855d","Type":"ContainerStarted","Data":"a16055db75b25076a62b6f358bf8c8eed9750e38d0a8658f4ebbe207795f1bb8"} Nov 24 17:24:08 crc kubenswrapper[4777]: I1124 17:24:08.669285 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 17:24:08 crc kubenswrapper[4777]: I1124 17:24:08.700542 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.365677171 podStartE2EDuration="6.700523026s" podCreationTimestamp="2025-11-24 17:24:02 +0000 UTC" firstStartedPulling="2025-11-24 17:24:03.551697202 +0000 UTC m=+1401.710892251" lastFinishedPulling="2025-11-24 17:24:07.886543057 +0000 UTC m=+1406.045738106" observedRunningTime="2025-11-24 17:24:08.691987399 +0000 UTC m=+1406.851182468" watchObservedRunningTime="2025-11-24 17:24:08.700523026 +0000 UTC m=+1406.859718065" Nov 24 17:24:08 crc kubenswrapper[4777]: I1124 17:24:08.701399 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-rjw7h" podStartSLOduration=6.701393761 podStartE2EDuration="6.701393761s" podCreationTimestamp="2025-11-24 17:24:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:24:04.632071684 +0000 UTC m=+1402.791266733" watchObservedRunningTime="2025-11-24 17:24:08.701393761 +0000 UTC m=+1406.860588810" Nov 24 17:24:09 crc kubenswrapper[4777]: I1124 17:24:09.679919 4777 generic.go:334] "Generic (PLEG): container finished" podID="2c1082e3-ae04-40f7-83ae-f7b7299b26b8" containerID="1b45413f849b6a67ffd7d2107e94d990e4106b4558326e9226a3e9c6a4c03f4d" exitCode=0 Nov 24 17:24:09 crc kubenswrapper[4777]: I1124 17:24:09.679992 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rjw7h" event={"ID":"2c1082e3-ae04-40f7-83ae-f7b7299b26b8","Type":"ContainerDied","Data":"1b45413f849b6a67ffd7d2107e94d990e4106b4558326e9226a3e9c6a4c03f4d"} Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.213879 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.341369 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmvnn\" (UniqueName: \"kubernetes.io/projected/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-kube-api-access-wmvnn\") pod \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.341631 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-config-data\") pod \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.341662 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-scripts\") pod \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.341704 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-combined-ca-bundle\") pod \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\" (UID: \"2c1082e3-ae04-40f7-83ae-f7b7299b26b8\") " Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.348145 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-scripts" (OuterVolumeSpecName: "scripts") pod "2c1082e3-ae04-40f7-83ae-f7b7299b26b8" (UID: "2c1082e3-ae04-40f7-83ae-f7b7299b26b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.348220 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-kube-api-access-wmvnn" (OuterVolumeSpecName: "kube-api-access-wmvnn") pod "2c1082e3-ae04-40f7-83ae-f7b7299b26b8" (UID: "2c1082e3-ae04-40f7-83ae-f7b7299b26b8"). InnerVolumeSpecName "kube-api-access-wmvnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.373863 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-config-data" (OuterVolumeSpecName: "config-data") pod "2c1082e3-ae04-40f7-83ae-f7b7299b26b8" (UID: "2c1082e3-ae04-40f7-83ae-f7b7299b26b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.382568 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c1082e3-ae04-40f7-83ae-f7b7299b26b8" (UID: "2c1082e3-ae04-40f7-83ae-f7b7299b26b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.444801 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.444835 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.444850 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.444864 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmvnn\" (UniqueName: \"kubernetes.io/projected/2c1082e3-ae04-40f7-83ae-f7b7299b26b8-kube-api-access-wmvnn\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.721295 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rjw7h" event={"ID":"2c1082e3-ae04-40f7-83ae-f7b7299b26b8","Type":"ContainerDied","Data":"cd3204032522c06f9d50f81f8e42aaeb8b9a791fb67cda5fc6469cbf2929e135"} Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.721701 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd3204032522c06f9d50f81f8e42aaeb8b9a791fb67cda5fc6469cbf2929e135" Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.721397 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rjw7h" Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.823520 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.823765 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" containerName="nova-api-log" containerID="cri-o://a0cd6a9f01d4e328ed595112a30fd680948d429e10c36688f79eddaaf8c213b6" gracePeriod=30 Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.823907 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" containerName="nova-api-api" containerID="cri-o://780deec4df91e72ee6fb8e88db0289a8bce65141a13ec2700f0c77e5d58a6068" gracePeriod=30 Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.904718 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.905016 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="5a9b3573-39b1-4336-9048-670f2ba47c23" containerName="nova-scheduler-scheduler" containerID="cri-o://24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c" gracePeriod=30 Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.918572 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.918815 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" containerName="nova-metadata-log" containerID="cri-o://f04868a0ddb0b3bf2ae3d10e99a7f026574673b7cf2a17a00218da9b22fa4e23" gracePeriod=30 Nov 24 17:24:11 crc kubenswrapper[4777]: I1124 17:24:11.919295 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" containerName="nova-metadata-metadata" containerID="cri-o://ad8e30e2fb3ae38dd466b51ed874992a582c1b29e9ef0571df393fd62f4276ea" gracePeriod=30 Nov 24 17:24:12 crc kubenswrapper[4777]: I1124 17:24:12.732810 4777 generic.go:334] "Generic (PLEG): container finished" podID="5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" containerID="a0cd6a9f01d4e328ed595112a30fd680948d429e10c36688f79eddaaf8c213b6" exitCode=143 Nov 24 17:24:12 crc kubenswrapper[4777]: I1124 17:24:12.732895 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e","Type":"ContainerDied","Data":"a0cd6a9f01d4e328ed595112a30fd680948d429e10c36688f79eddaaf8c213b6"} Nov 24 17:24:12 crc kubenswrapper[4777]: I1124 17:24:12.736847 4777 generic.go:334] "Generic (PLEG): container finished" podID="ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" containerID="f04868a0ddb0b3bf2ae3d10e99a7f026574673b7cf2a17a00218da9b22fa4e23" exitCode=143 Nov 24 17:24:12 crc kubenswrapper[4777]: I1124 17:24:12.736888 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf","Type":"ContainerDied","Data":"f04868a0ddb0b3bf2ae3d10e99a7f026574673b7cf2a17a00218da9b22fa4e23"} Nov 24 17:24:13 crc kubenswrapper[4777]: E1124 17:24:13.441003 4777 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 17:24:13 crc kubenswrapper[4777]: E1124 17:24:13.442681 4777 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 17:24:13 crc kubenswrapper[4777]: E1124 17:24:13.443955 4777 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 17:24:13 crc kubenswrapper[4777]: E1124 17:24:13.444082 4777 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="5a9b3573-39b1-4336-9048-670f2ba47c23" containerName="nova-scheduler-scheduler" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.337327 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.217:8775/\": read tcp 10.217.0.2:54590->10.217.0.217:8775: read: connection reset by peer" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.337318 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.217:8775/\": read tcp 10.217.0.2:54598->10.217.0.217:8775: read: connection reset by peer" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.729422 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.774421 4777 generic.go:334] "Generic (PLEG): container finished" podID="ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" containerID="ad8e30e2fb3ae38dd466b51ed874992a582c1b29e9ef0571df393fd62f4276ea" exitCode=0 Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.774461 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf","Type":"ContainerDied","Data":"ad8e30e2fb3ae38dd466b51ed874992a582c1b29e9ef0571df393fd62f4276ea"} Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.777224 4777 generic.go:334] "Generic (PLEG): container finished" podID="5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" containerID="780deec4df91e72ee6fb8e88db0289a8bce65141a13ec2700f0c77e5d58a6068" exitCode=0 Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.777268 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e","Type":"ContainerDied","Data":"780deec4df91e72ee6fb8e88db0289a8bce65141a13ec2700f0c77e5d58a6068"} Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.777297 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.777311 4777 scope.go:117] "RemoveContainer" containerID="780deec4df91e72ee6fb8e88db0289a8bce65141a13ec2700f0c77e5d58a6068" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.777298 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e","Type":"ContainerDied","Data":"af9cf68ef7847f2dc8e00ea29beb884011222ba9a53c420406cfbaeca88be471"} Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.809050 4777 scope.go:117] "RemoveContainer" containerID="a0cd6a9f01d4e328ed595112a30fd680948d429e10c36688f79eddaaf8c213b6" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.827868 4777 scope.go:117] "RemoveContainer" containerID="780deec4df91e72ee6fb8e88db0289a8bce65141a13ec2700f0c77e5d58a6068" Nov 24 17:24:15 crc kubenswrapper[4777]: E1124 17:24:15.828263 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"780deec4df91e72ee6fb8e88db0289a8bce65141a13ec2700f0c77e5d58a6068\": container with ID starting with 780deec4df91e72ee6fb8e88db0289a8bce65141a13ec2700f0c77e5d58a6068 not found: ID does not exist" containerID="780deec4df91e72ee6fb8e88db0289a8bce65141a13ec2700f0c77e5d58a6068" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.828324 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"780deec4df91e72ee6fb8e88db0289a8bce65141a13ec2700f0c77e5d58a6068"} err="failed to get container status \"780deec4df91e72ee6fb8e88db0289a8bce65141a13ec2700f0c77e5d58a6068\": rpc error: code = NotFound desc = could not find container \"780deec4df91e72ee6fb8e88db0289a8bce65141a13ec2700f0c77e5d58a6068\": container with ID starting with 780deec4df91e72ee6fb8e88db0289a8bce65141a13ec2700f0c77e5d58a6068 not found: ID does not exist" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.828354 4777 scope.go:117] "RemoveContainer" containerID="a0cd6a9f01d4e328ed595112a30fd680948d429e10c36688f79eddaaf8c213b6" Nov 24 17:24:15 crc kubenswrapper[4777]: E1124 17:24:15.828800 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0cd6a9f01d4e328ed595112a30fd680948d429e10c36688f79eddaaf8c213b6\": container with ID starting with a0cd6a9f01d4e328ed595112a30fd680948d429e10c36688f79eddaaf8c213b6 not found: ID does not exist" containerID="a0cd6a9f01d4e328ed595112a30fd680948d429e10c36688f79eddaaf8c213b6" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.828851 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0cd6a9f01d4e328ed595112a30fd680948d429e10c36688f79eddaaf8c213b6"} err="failed to get container status \"a0cd6a9f01d4e328ed595112a30fd680948d429e10c36688f79eddaaf8c213b6\": rpc error: code = NotFound desc = could not find container \"a0cd6a9f01d4e328ed595112a30fd680948d429e10c36688f79eddaaf8c213b6\": container with ID starting with a0cd6a9f01d4e328ed595112a30fd680948d429e10c36688f79eddaaf8c213b6 not found: ID does not exist" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.840731 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-config-data\") pod \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.841263 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd2bc\" (UniqueName: \"kubernetes.io/projected/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-kube-api-access-cd2bc\") pod \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.841439 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-public-tls-certs\") pod \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.841635 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-combined-ca-bundle\") pod \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.842313 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-logs\") pod \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.842539 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-internal-tls-certs\") pod \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\" (UID: \"5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e\") " Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.844746 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-logs" (OuterVolumeSpecName: "logs") pod "5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" (UID: "5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.848161 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-kube-api-access-cd2bc" (OuterVolumeSpecName: "kube-api-access-cd2bc") pod "5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" (UID: "5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e"). InnerVolumeSpecName "kube-api-access-cd2bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.877997 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-config-data" (OuterVolumeSpecName: "config-data") pod "5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" (UID: "5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.878633 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" (UID: "5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.902038 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" (UID: "5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.902669 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.914406 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" (UID: "5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.946510 4777 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.946545 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.946557 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd2bc\" (UniqueName: \"kubernetes.io/projected/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-kube-api-access-cd2bc\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.946567 4777 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.946577 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:15 crc kubenswrapper[4777]: I1124 17:24:15.946588 4777 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.048159 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-nova-metadata-tls-certs\") pod \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.048200 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-config-data\") pod \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.048273 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-logs\") pod \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.048294 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ts725\" (UniqueName: \"kubernetes.io/projected/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-kube-api-access-ts725\") pod \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.048373 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-combined-ca-bundle\") pod \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\" (UID: \"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf\") " Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.048772 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-logs" (OuterVolumeSpecName: "logs") pod "ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" (UID: "ab9c09ec-00a0-4bae-a64b-08e9fb689bcf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.049080 4777 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.054011 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-kube-api-access-ts725" (OuterVolumeSpecName: "kube-api-access-ts725") pod "ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" (UID: "ab9c09ec-00a0-4bae-a64b-08e9fb689bcf"). InnerVolumeSpecName "kube-api-access-ts725". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.076227 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-config-data" (OuterVolumeSpecName: "config-data") pod "ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" (UID: "ab9c09ec-00a0-4bae-a64b-08e9fb689bcf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.089894 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" (UID: "ab9c09ec-00a0-4bae-a64b-08e9fb689bcf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.105682 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" (UID: "ab9c09ec-00a0-4bae-a64b-08e9fb689bcf"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.151348 4777 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.151391 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.151406 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ts725\" (UniqueName: \"kubernetes.io/projected/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-kube-api-access-ts725\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.151421 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.226763 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.253193 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.267017 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 17:24:16 crc kubenswrapper[4777]: E1124 17:24:16.267588 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" containerName="nova-metadata-metadata" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.267605 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" containerName="nova-metadata-metadata" Nov 24 17:24:16 crc kubenswrapper[4777]: E1124 17:24:16.267622 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c1082e3-ae04-40f7-83ae-f7b7299b26b8" containerName="nova-manage" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.267632 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c1082e3-ae04-40f7-83ae-f7b7299b26b8" containerName="nova-manage" Nov 24 17:24:16 crc kubenswrapper[4777]: E1124 17:24:16.267652 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" containerName="nova-api-api" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.267660 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" containerName="nova-api-api" Nov 24 17:24:16 crc kubenswrapper[4777]: E1124 17:24:16.267670 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" containerName="nova-metadata-log" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.267678 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" containerName="nova-metadata-log" Nov 24 17:24:16 crc kubenswrapper[4777]: E1124 17:24:16.267695 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" containerName="nova-api-log" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.267702 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" containerName="nova-api-log" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.267944 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" containerName="nova-api-api" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.267985 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" containerName="nova-api-log" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.268007 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c1082e3-ae04-40f7-83ae-f7b7299b26b8" containerName="nova-manage" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.268026 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" containerName="nova-metadata-log" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.268037 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" containerName="nova-metadata-metadata" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.269466 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.271657 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.274219 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.276585 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.285139 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.356819 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f474cf0f-37fd-495d-ac31-493641c222a7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.356883 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f474cf0f-37fd-495d-ac31-493641c222a7-public-tls-certs\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.357020 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f474cf0f-37fd-495d-ac31-493641c222a7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.357161 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f474cf0f-37fd-495d-ac31-493641c222a7-config-data\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.357196 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f474cf0f-37fd-495d-ac31-493641c222a7-logs\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.357267 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nsnp\" (UniqueName: \"kubernetes.io/projected/f474cf0f-37fd-495d-ac31-493641c222a7-kube-api-access-9nsnp\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.459258 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nsnp\" (UniqueName: \"kubernetes.io/projected/f474cf0f-37fd-495d-ac31-493641c222a7-kube-api-access-9nsnp\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.459317 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f474cf0f-37fd-495d-ac31-493641c222a7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.459367 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f474cf0f-37fd-495d-ac31-493641c222a7-public-tls-certs\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.459416 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f474cf0f-37fd-495d-ac31-493641c222a7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.459538 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f474cf0f-37fd-495d-ac31-493641c222a7-config-data\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.459582 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f474cf0f-37fd-495d-ac31-493641c222a7-logs\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.460196 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f474cf0f-37fd-495d-ac31-493641c222a7-logs\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.462688 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f474cf0f-37fd-495d-ac31-493641c222a7-public-tls-certs\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.462796 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f474cf0f-37fd-495d-ac31-493641c222a7-config-data\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.463267 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f474cf0f-37fd-495d-ac31-493641c222a7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.463588 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f474cf0f-37fd-495d-ac31-493641c222a7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.484604 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nsnp\" (UniqueName: \"kubernetes.io/projected/f474cf0f-37fd-495d-ac31-493641c222a7-kube-api-access-9nsnp\") pod \"nova-api-0\" (UID: \"f474cf0f-37fd-495d-ac31-493641c222a7\") " pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.588023 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.793591 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab9c09ec-00a0-4bae-a64b-08e9fb689bcf","Type":"ContainerDied","Data":"ce9699c18e2d659d0691a72b36cb9b9fa1fc8016bb37120187f123eae898da76"} Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.793950 4777 scope.go:117] "RemoveContainer" containerID="ad8e30e2fb3ae38dd466b51ed874992a582c1b29e9ef0571df393fd62f4276ea" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.793791 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.871187 4777 scope.go:117] "RemoveContainer" containerID="f04868a0ddb0b3bf2ae3d10e99a7f026574673b7cf2a17a00218da9b22fa4e23" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.890795 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.917506 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.927494 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.929733 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.934456 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.934796 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 17:24:16 crc kubenswrapper[4777]: I1124 17:24:16.936165 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.046461 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.091943 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4da8a01e-98cb-4d18-b86e-c281010dcf8e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4da8a01e-98cb-4d18-b86e-c281010dcf8e\") " pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.092185 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4da8a01e-98cb-4d18-b86e-c281010dcf8e-logs\") pod \"nova-metadata-0\" (UID: \"4da8a01e-98cb-4d18-b86e-c281010dcf8e\") " pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.092281 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4da8a01e-98cb-4d18-b86e-c281010dcf8e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4da8a01e-98cb-4d18-b86e-c281010dcf8e\") " pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.092570 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lph8h\" (UniqueName: \"kubernetes.io/projected/4da8a01e-98cb-4d18-b86e-c281010dcf8e-kube-api-access-lph8h\") pod \"nova-metadata-0\" (UID: \"4da8a01e-98cb-4d18-b86e-c281010dcf8e\") " pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.092681 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4da8a01e-98cb-4d18-b86e-c281010dcf8e-config-data\") pod \"nova-metadata-0\" (UID: \"4da8a01e-98cb-4d18-b86e-c281010dcf8e\") " pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.194644 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lph8h\" (UniqueName: \"kubernetes.io/projected/4da8a01e-98cb-4d18-b86e-c281010dcf8e-kube-api-access-lph8h\") pod \"nova-metadata-0\" (UID: \"4da8a01e-98cb-4d18-b86e-c281010dcf8e\") " pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.194704 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4da8a01e-98cb-4d18-b86e-c281010dcf8e-config-data\") pod \"nova-metadata-0\" (UID: \"4da8a01e-98cb-4d18-b86e-c281010dcf8e\") " pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.194753 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4da8a01e-98cb-4d18-b86e-c281010dcf8e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4da8a01e-98cb-4d18-b86e-c281010dcf8e\") " pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.194832 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4da8a01e-98cb-4d18-b86e-c281010dcf8e-logs\") pod \"nova-metadata-0\" (UID: \"4da8a01e-98cb-4d18-b86e-c281010dcf8e\") " pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.194869 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4da8a01e-98cb-4d18-b86e-c281010dcf8e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4da8a01e-98cb-4d18-b86e-c281010dcf8e\") " pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.195928 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4da8a01e-98cb-4d18-b86e-c281010dcf8e-logs\") pod \"nova-metadata-0\" (UID: \"4da8a01e-98cb-4d18-b86e-c281010dcf8e\") " pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.200762 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4da8a01e-98cb-4d18-b86e-c281010dcf8e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4da8a01e-98cb-4d18-b86e-c281010dcf8e\") " pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.201257 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4da8a01e-98cb-4d18-b86e-c281010dcf8e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4da8a01e-98cb-4d18-b86e-c281010dcf8e\") " pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.202936 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4da8a01e-98cb-4d18-b86e-c281010dcf8e-config-data\") pod \"nova-metadata-0\" (UID: \"4da8a01e-98cb-4d18-b86e-c281010dcf8e\") " pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.221722 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lph8h\" (UniqueName: \"kubernetes.io/projected/4da8a01e-98cb-4d18-b86e-c281010dcf8e-kube-api-access-lph8h\") pod \"nova-metadata-0\" (UID: \"4da8a01e-98cb-4d18-b86e-c281010dcf8e\") " pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.263494 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e" path="/var/lib/kubelet/pods/5b0fb0ea-87f0-41b1-b3e7-1492646c9d5e/volumes" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.264268 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab9c09ec-00a0-4bae-a64b-08e9fb689bcf" path="/var/lib/kubelet/pods/ab9c09ec-00a0-4bae-a64b-08e9fb689bcf/volumes" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.267175 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.600319 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.703449 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bldlw\" (UniqueName: \"kubernetes.io/projected/5a9b3573-39b1-4336-9048-670f2ba47c23-kube-api-access-bldlw\") pod \"5a9b3573-39b1-4336-9048-670f2ba47c23\" (UID: \"5a9b3573-39b1-4336-9048-670f2ba47c23\") " Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.703495 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9b3573-39b1-4336-9048-670f2ba47c23-config-data\") pod \"5a9b3573-39b1-4336-9048-670f2ba47c23\" (UID: \"5a9b3573-39b1-4336-9048-670f2ba47c23\") " Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.703658 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9b3573-39b1-4336-9048-670f2ba47c23-combined-ca-bundle\") pod \"5a9b3573-39b1-4336-9048-670f2ba47c23\" (UID: \"5a9b3573-39b1-4336-9048-670f2ba47c23\") " Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.712182 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a9b3573-39b1-4336-9048-670f2ba47c23-kube-api-access-bldlw" (OuterVolumeSpecName: "kube-api-access-bldlw") pod "5a9b3573-39b1-4336-9048-670f2ba47c23" (UID: "5a9b3573-39b1-4336-9048-670f2ba47c23"). InnerVolumeSpecName "kube-api-access-bldlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.795147 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a9b3573-39b1-4336-9048-670f2ba47c23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a9b3573-39b1-4336-9048-670f2ba47c23" (UID: "5a9b3573-39b1-4336-9048-670f2ba47c23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.806080 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bldlw\" (UniqueName: \"kubernetes.io/projected/5a9b3573-39b1-4336-9048-670f2ba47c23-kube-api-access-bldlw\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.806102 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9b3573-39b1-4336-9048-670f2ba47c23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.809155 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a9b3573-39b1-4336-9048-670f2ba47c23-config-data" (OuterVolumeSpecName: "config-data") pod "5a9b3573-39b1-4336-9048-670f2ba47c23" (UID: "5a9b3573-39b1-4336-9048-670f2ba47c23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.822648 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f474cf0f-37fd-495d-ac31-493641c222a7","Type":"ContainerStarted","Data":"72d3ffe0a8185713bdb32be7e27701f29c762bd669d437105b519a3cede10a3a"} Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.822700 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f474cf0f-37fd-495d-ac31-493641c222a7","Type":"ContainerStarted","Data":"0821d2ca1612d36151cc3316f41666e194aacdb02a289624c799bf34f83ac420"} Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.828471 4777 generic.go:334] "Generic (PLEG): container finished" podID="5a9b3573-39b1-4336-9048-670f2ba47c23" containerID="24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c" exitCode=0 Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.828597 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.830380 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a9b3573-39b1-4336-9048-670f2ba47c23","Type":"ContainerDied","Data":"24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c"} Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.830420 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a9b3573-39b1-4336-9048-670f2ba47c23","Type":"ContainerDied","Data":"e31180a6f104245a6d82773a50030c76e48fdb34ae48d193a79cacce67d35746"} Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.830440 4777 scope.go:117] "RemoveContainer" containerID="24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.851631 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 17:24:17 crc kubenswrapper[4777]: W1124 17:24:17.876178 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4da8a01e_98cb_4d18_b86e_c281010dcf8e.slice/crio-2f24bf1a819107cf5af1acaf98f8e15fa88b4aecd9015469ac98d3877273b9ca WatchSource:0}: Error finding container 2f24bf1a819107cf5af1acaf98f8e15fa88b4aecd9015469ac98d3877273b9ca: Status 404 returned error can't find the container with id 2f24bf1a819107cf5af1acaf98f8e15fa88b4aecd9015469ac98d3877273b9ca Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.898141 4777 scope.go:117] "RemoveContainer" containerID="24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c" Nov 24 17:24:17 crc kubenswrapper[4777]: E1124 17:24:17.902180 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c\": container with ID starting with 24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c not found: ID does not exist" containerID="24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.902226 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c"} err="failed to get container status \"24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c\": rpc error: code = NotFound desc = could not find container \"24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c\": container with ID starting with 24f5029c906c1c04c5707fda9efd0e200076d0945607cd916bccf74313ada02c not found: ID does not exist" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.907667 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9b3573-39b1-4336-9048-670f2ba47c23-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.911196 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.923388 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.957469 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:24:17 crc kubenswrapper[4777]: E1124 17:24:17.957928 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9b3573-39b1-4336-9048-670f2ba47c23" containerName="nova-scheduler-scheduler" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.957943 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9b3573-39b1-4336-9048-670f2ba47c23" containerName="nova-scheduler-scheduler" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.958186 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9b3573-39b1-4336-9048-670f2ba47c23" containerName="nova-scheduler-scheduler" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.958914 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.962419 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 17:24:17 crc kubenswrapper[4777]: I1124 17:24:17.966992 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.111206 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a63c908-3ee5-449c-ab1c-6ca209338bee-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2a63c908-3ee5-449c-ab1c-6ca209338bee\") " pod="openstack/nova-scheduler-0" Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.111654 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rvb2\" (UniqueName: \"kubernetes.io/projected/2a63c908-3ee5-449c-ab1c-6ca209338bee-kube-api-access-2rvb2\") pod \"nova-scheduler-0\" (UID: \"2a63c908-3ee5-449c-ab1c-6ca209338bee\") " pod="openstack/nova-scheduler-0" Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.111784 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a63c908-3ee5-449c-ab1c-6ca209338bee-config-data\") pod \"nova-scheduler-0\" (UID: \"2a63c908-3ee5-449c-ab1c-6ca209338bee\") " pod="openstack/nova-scheduler-0" Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.213463 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rvb2\" (UniqueName: \"kubernetes.io/projected/2a63c908-3ee5-449c-ab1c-6ca209338bee-kube-api-access-2rvb2\") pod \"nova-scheduler-0\" (UID: \"2a63c908-3ee5-449c-ab1c-6ca209338bee\") " pod="openstack/nova-scheduler-0" Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.213564 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a63c908-3ee5-449c-ab1c-6ca209338bee-config-data\") pod \"nova-scheduler-0\" (UID: \"2a63c908-3ee5-449c-ab1c-6ca209338bee\") " pod="openstack/nova-scheduler-0" Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.213707 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a63c908-3ee5-449c-ab1c-6ca209338bee-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2a63c908-3ee5-449c-ab1c-6ca209338bee\") " pod="openstack/nova-scheduler-0" Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.221775 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a63c908-3ee5-449c-ab1c-6ca209338bee-config-data\") pod \"nova-scheduler-0\" (UID: \"2a63c908-3ee5-449c-ab1c-6ca209338bee\") " pod="openstack/nova-scheduler-0" Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.221778 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a63c908-3ee5-449c-ab1c-6ca209338bee-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2a63c908-3ee5-449c-ab1c-6ca209338bee\") " pod="openstack/nova-scheduler-0" Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.234524 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rvb2\" (UniqueName: \"kubernetes.io/projected/2a63c908-3ee5-449c-ab1c-6ca209338bee-kube-api-access-2rvb2\") pod \"nova-scheduler-0\" (UID: \"2a63c908-3ee5-449c-ab1c-6ca209338bee\") " pod="openstack/nova-scheduler-0" Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.278779 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.782526 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 17:24:18 crc kubenswrapper[4777]: W1124 17:24:18.785352 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a63c908_3ee5_449c_ab1c_6ca209338bee.slice/crio-58c62c90f5c7ac807347532899bd7b9a3417a2b0d2f727353c8c22a3b20706ca WatchSource:0}: Error finding container 58c62c90f5c7ac807347532899bd7b9a3417a2b0d2f727353c8c22a3b20706ca: Status 404 returned error can't find the container with id 58c62c90f5c7ac807347532899bd7b9a3417a2b0d2f727353c8c22a3b20706ca Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.860244 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4da8a01e-98cb-4d18-b86e-c281010dcf8e","Type":"ContainerStarted","Data":"6ec96106c0a1c28f4057d67b837155e139cbf8b733539b3f6ffd613698be1cc2"} Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.860308 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4da8a01e-98cb-4d18-b86e-c281010dcf8e","Type":"ContainerStarted","Data":"b9e9f15232630f54c48208c83e0b8063109fde31dce1c7571ba474d14a4618f3"} Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.860327 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4da8a01e-98cb-4d18-b86e-c281010dcf8e","Type":"ContainerStarted","Data":"2f24bf1a819107cf5af1acaf98f8e15fa88b4aecd9015469ac98d3877273b9ca"} Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.863825 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2a63c908-3ee5-449c-ab1c-6ca209338bee","Type":"ContainerStarted","Data":"58c62c90f5c7ac807347532899bd7b9a3417a2b0d2f727353c8c22a3b20706ca"} Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.869851 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f474cf0f-37fd-495d-ac31-493641c222a7","Type":"ContainerStarted","Data":"5c7c2000d2dafd80ffc7cca2f2a9906f357525bdfe31761e36de290bbdc79c2e"} Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.905664 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.905641496 podStartE2EDuration="2.905641496s" podCreationTimestamp="2025-11-24 17:24:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:24:18.903180865 +0000 UTC m=+1417.062375924" watchObservedRunningTime="2025-11-24 17:24:18.905641496 +0000 UTC m=+1417.064836545" Nov 24 17:24:18 crc kubenswrapper[4777]: I1124 17:24:18.915101 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.9150823790000002 podStartE2EDuration="2.915082379s" podCreationTimestamp="2025-11-24 17:24:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:24:18.882322332 +0000 UTC m=+1417.041517381" watchObservedRunningTime="2025-11-24 17:24:18.915082379 +0000 UTC m=+1417.074277428" Nov 24 17:24:19 crc kubenswrapper[4777]: I1124 17:24:19.259208 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a9b3573-39b1-4336-9048-670f2ba47c23" path="/var/lib/kubelet/pods/5a9b3573-39b1-4336-9048-670f2ba47c23/volumes" Nov 24 17:24:19 crc kubenswrapper[4777]: I1124 17:24:19.893830 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2a63c908-3ee5-449c-ab1c-6ca209338bee","Type":"ContainerStarted","Data":"3cb2522ab46c20ca57edba6ccd809b6521ee758bd2047132c6a3b973703d5080"} Nov 24 17:24:19 crc kubenswrapper[4777]: I1124 17:24:19.911541 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.9115259140000003 podStartE2EDuration="2.911525914s" podCreationTimestamp="2025-11-24 17:24:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:24:19.910668319 +0000 UTC m=+1418.069863398" watchObservedRunningTime="2025-11-24 17:24:19.911525914 +0000 UTC m=+1418.070720963" Nov 24 17:24:22 crc kubenswrapper[4777]: I1124 17:24:22.267595 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 17:24:22 crc kubenswrapper[4777]: I1124 17:24:22.268234 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 17:24:23 crc kubenswrapper[4777]: I1124 17:24:23.279610 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 17:24:26 crc kubenswrapper[4777]: I1124 17:24:26.589577 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 17:24:26 crc kubenswrapper[4777]: I1124 17:24:26.590039 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 17:24:27 crc kubenswrapper[4777]: I1124 17:24:27.267771 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 17:24:27 crc kubenswrapper[4777]: I1124 17:24:27.267813 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 17:24:27 crc kubenswrapper[4777]: I1124 17:24:27.603123 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f474cf0f-37fd-495d-ac31-493641c222a7" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.228:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 17:24:27 crc kubenswrapper[4777]: I1124 17:24:27.603177 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f474cf0f-37fd-495d-ac31-493641c222a7" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.228:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 17:24:28 crc kubenswrapper[4777]: I1124 17:24:28.279260 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4da8a01e-98cb-4d18-b86e-c281010dcf8e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.229:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 17:24:28 crc kubenswrapper[4777]: I1124 17:24:28.280223 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4da8a01e-98cb-4d18-b86e-c281010dcf8e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.229:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 17:24:28 crc kubenswrapper[4777]: I1124 17:24:28.279394 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 17:24:28 crc kubenswrapper[4777]: I1124 17:24:28.318940 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 17:24:28 crc kubenswrapper[4777]: I1124 17:24:28.364675 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 17:24:33 crc kubenswrapper[4777]: I1124 17:24:33.039617 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 17:24:36 crc kubenswrapper[4777]: I1124 17:24:36.597670 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 17:24:36 crc kubenswrapper[4777]: I1124 17:24:36.598601 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 17:24:36 crc kubenswrapper[4777]: I1124 17:24:36.598938 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 17:24:36 crc kubenswrapper[4777]: I1124 17:24:36.599025 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 17:24:36 crc kubenswrapper[4777]: I1124 17:24:36.608405 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 17:24:36 crc kubenswrapper[4777]: I1124 17:24:36.613159 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 17:24:37 crc kubenswrapper[4777]: I1124 17:24:37.273661 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 17:24:37 crc kubenswrapper[4777]: I1124 17:24:37.277134 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 17:24:37 crc kubenswrapper[4777]: I1124 17:24:37.289577 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 17:24:37 crc kubenswrapper[4777]: I1124 17:24:37.421768 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.444822 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-sync-57hm5"] Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.457825 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-sync-57hm5"] Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.556229 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-sync-9d788"] Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.557500 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.561417 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.569845 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-9d788"] Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.594455 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt5s2\" (UniqueName: \"kubernetes.io/projected/5e74119f-b07b-4531-b5fa-5a92a9c6625b-kube-api-access-rt5s2\") pod \"cloudkitty-db-sync-9d788\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.594516 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/5e74119f-b07b-4531-b5fa-5a92a9c6625b-certs\") pod \"cloudkitty-db-sync-9d788\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.594651 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-scripts\") pod \"cloudkitty-db-sync-9d788\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.594696 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-config-data\") pod \"cloudkitty-db-sync-9d788\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.594721 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-combined-ca-bundle\") pod \"cloudkitty-db-sync-9d788\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.696123 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt5s2\" (UniqueName: \"kubernetes.io/projected/5e74119f-b07b-4531-b5fa-5a92a9c6625b-kube-api-access-rt5s2\") pod \"cloudkitty-db-sync-9d788\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.696172 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/5e74119f-b07b-4531-b5fa-5a92a9c6625b-certs\") pod \"cloudkitty-db-sync-9d788\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.696231 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-scripts\") pod \"cloudkitty-db-sync-9d788\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.696256 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-config-data\") pod \"cloudkitty-db-sync-9d788\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.696279 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-combined-ca-bundle\") pod \"cloudkitty-db-sync-9d788\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.702920 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-scripts\") pod \"cloudkitty-db-sync-9d788\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.703727 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-combined-ca-bundle\") pod \"cloudkitty-db-sync-9d788\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.704427 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/5e74119f-b07b-4531-b5fa-5a92a9c6625b-certs\") pod \"cloudkitty-db-sync-9d788\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.710264 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-config-data\") pod \"cloudkitty-db-sync-9d788\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.720592 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt5s2\" (UniqueName: \"kubernetes.io/projected/5e74119f-b07b-4531-b5fa-5a92a9c6625b-kube-api-access-rt5s2\") pod \"cloudkitty-db-sync-9d788\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.739095 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h2fbd"] Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.741444 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.754616 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h2fbd"] Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.812503 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kvj5\" (UniqueName: \"kubernetes.io/projected/e36159e1-7be5-49a7-8417-57253b4e7be0-kube-api-access-7kvj5\") pod \"redhat-operators-h2fbd\" (UID: \"e36159e1-7be5-49a7-8417-57253b4e7be0\") " pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.812731 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e36159e1-7be5-49a7-8417-57253b4e7be0-utilities\") pod \"redhat-operators-h2fbd\" (UID: \"e36159e1-7be5-49a7-8417-57253b4e7be0\") " pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.813347 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e36159e1-7be5-49a7-8417-57253b4e7be0-catalog-content\") pod \"redhat-operators-h2fbd\" (UID: \"e36159e1-7be5-49a7-8417-57253b4e7be0\") " pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.877079 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.915275 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kvj5\" (UniqueName: \"kubernetes.io/projected/e36159e1-7be5-49a7-8417-57253b4e7be0-kube-api-access-7kvj5\") pod \"redhat-operators-h2fbd\" (UID: \"e36159e1-7be5-49a7-8417-57253b4e7be0\") " pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.915418 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e36159e1-7be5-49a7-8417-57253b4e7be0-utilities\") pod \"redhat-operators-h2fbd\" (UID: \"e36159e1-7be5-49a7-8417-57253b4e7be0\") " pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.915657 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e36159e1-7be5-49a7-8417-57253b4e7be0-catalog-content\") pod \"redhat-operators-h2fbd\" (UID: \"e36159e1-7be5-49a7-8417-57253b4e7be0\") " pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.916222 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e36159e1-7be5-49a7-8417-57253b4e7be0-utilities\") pod \"redhat-operators-h2fbd\" (UID: \"e36159e1-7be5-49a7-8417-57253b4e7be0\") " pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.916399 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e36159e1-7be5-49a7-8417-57253b4e7be0-catalog-content\") pod \"redhat-operators-h2fbd\" (UID: \"e36159e1-7be5-49a7-8417-57253b4e7be0\") " pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:24:47 crc kubenswrapper[4777]: I1124 17:24:47.937310 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kvj5\" (UniqueName: \"kubernetes.io/projected/e36159e1-7be5-49a7-8417-57253b4e7be0-kube-api-access-7kvj5\") pod \"redhat-operators-h2fbd\" (UID: \"e36159e1-7be5-49a7-8417-57253b4e7be0\") " pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:24:48 crc kubenswrapper[4777]: I1124 17:24:48.152008 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:24:48 crc kubenswrapper[4777]: I1124 17:24:48.382864 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-9d788"] Nov 24 17:24:48 crc kubenswrapper[4777]: I1124 17:24:48.566276 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-9d788" event={"ID":"5e74119f-b07b-4531-b5fa-5a92a9c6625b","Type":"ContainerStarted","Data":"df435a6c8e65906c43b3e360cc272386ddd688ae3c52bc6c8ceb68fadd335047"} Nov 24 17:24:49 crc kubenswrapper[4777]: I1124 17:24:49.255176 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5ad0d8c-d098-4a37-a8f3-2563fff8b620" path="/var/lib/kubelet/pods/a5ad0d8c-d098-4a37-a8f3-2563fff8b620/volumes" Nov 24 17:24:49 crc kubenswrapper[4777]: W1124 17:24:49.272463 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode36159e1_7be5_49a7_8417_57253b4e7be0.slice/crio-a0f2954607e8bcf3f1a5629142d5a65b7e1d0786518f39b33fef13d59889f470 WatchSource:0}: Error finding container a0f2954607e8bcf3f1a5629142d5a65b7e1d0786518f39b33fef13d59889f470: Status 404 returned error can't find the container with id a0f2954607e8bcf3f1a5629142d5a65b7e1d0786518f39b33fef13d59889f470 Nov 24 17:24:49 crc kubenswrapper[4777]: I1124 17:24:49.284867 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h2fbd"] Nov 24 17:24:49 crc kubenswrapper[4777]: I1124 17:24:49.579742 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h2fbd" event={"ID":"e36159e1-7be5-49a7-8417-57253b4e7be0","Type":"ContainerStarted","Data":"ab7e9e98a34b1988684706f4ddeaf816d942278c54ee4a8d6a45503c68ca6d2c"} Nov 24 17:24:49 crc kubenswrapper[4777]: I1124 17:24:49.580064 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h2fbd" event={"ID":"e36159e1-7be5-49a7-8417-57253b4e7be0","Type":"ContainerStarted","Data":"a0f2954607e8bcf3f1a5629142d5a65b7e1d0786518f39b33fef13d59889f470"} Nov 24 17:24:49 crc kubenswrapper[4777]: I1124 17:24:49.582285 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-9d788" event={"ID":"5e74119f-b07b-4531-b5fa-5a92a9c6625b","Type":"ContainerStarted","Data":"828c4ec9b88e85872599f26f0bc8dbf67daad7b0252833a2b91593572ab5062b"} Nov 24 17:24:49 crc kubenswrapper[4777]: I1124 17:24:49.626187 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-sync-9d788" podStartSLOduration=2.463957878 podStartE2EDuration="2.626164052s" podCreationTimestamp="2025-11-24 17:24:47 +0000 UTC" firstStartedPulling="2025-11-24 17:24:48.396099123 +0000 UTC m=+1446.555294172" lastFinishedPulling="2025-11-24 17:24:48.558305297 +0000 UTC m=+1446.717500346" observedRunningTime="2025-11-24 17:24:49.617527632 +0000 UTC m=+1447.776722681" watchObservedRunningTime="2025-11-24 17:24:49.626164052 +0000 UTC m=+1447.785359101" Nov 24 17:24:49 crc kubenswrapper[4777]: I1124 17:24:49.752022 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:24:50 crc kubenswrapper[4777]: I1124 17:24:50.049400 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:24:50 crc kubenswrapper[4777]: I1124 17:24:50.049727 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="ceilometer-central-agent" containerID="cri-o://f5bc687fc5668ac4935952f9e02557d75c41fa5a98550fa1786c9d926584e7ae" gracePeriod=30 Nov 24 17:24:50 crc kubenswrapper[4777]: I1124 17:24:50.049799 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="proxy-httpd" containerID="cri-o://a16055db75b25076a62b6f358bf8c8eed9750e38d0a8658f4ebbe207795f1bb8" gracePeriod=30 Nov 24 17:24:50 crc kubenswrapper[4777]: I1124 17:24:50.049833 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="sg-core" containerID="cri-o://1a8c50da8aeb8068e5d4150d334c3b86a8ba7c7a17681cd61776dfe674563d98" gracePeriod=30 Nov 24 17:24:50 crc kubenswrapper[4777]: I1124 17:24:50.049875 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="ceilometer-notification-agent" containerID="cri-o://117c081d4641d4be6b7400e1d344b0278b56d84856124fa09bc36ccec9a2698e" gracePeriod=30 Nov 24 17:24:50 crc kubenswrapper[4777]: I1124 17:24:50.595894 4777 generic.go:334] "Generic (PLEG): container finished" podID="9412f05c-2d23-4124-abde-2f573bb6855d" containerID="a16055db75b25076a62b6f358bf8c8eed9750e38d0a8658f4ebbe207795f1bb8" exitCode=0 Nov 24 17:24:50 crc kubenswrapper[4777]: I1124 17:24:50.596257 4777 generic.go:334] "Generic (PLEG): container finished" podID="9412f05c-2d23-4124-abde-2f573bb6855d" containerID="1a8c50da8aeb8068e5d4150d334c3b86a8ba7c7a17681cd61776dfe674563d98" exitCode=2 Nov 24 17:24:50 crc kubenswrapper[4777]: I1124 17:24:50.596273 4777 generic.go:334] "Generic (PLEG): container finished" podID="9412f05c-2d23-4124-abde-2f573bb6855d" containerID="f5bc687fc5668ac4935952f9e02557d75c41fa5a98550fa1786c9d926584e7ae" exitCode=0 Nov 24 17:24:50 crc kubenswrapper[4777]: I1124 17:24:50.595982 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9412f05c-2d23-4124-abde-2f573bb6855d","Type":"ContainerDied","Data":"a16055db75b25076a62b6f358bf8c8eed9750e38d0a8658f4ebbe207795f1bb8"} Nov 24 17:24:50 crc kubenswrapper[4777]: I1124 17:24:50.596360 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9412f05c-2d23-4124-abde-2f573bb6855d","Type":"ContainerDied","Data":"1a8c50da8aeb8068e5d4150d334c3b86a8ba7c7a17681cd61776dfe674563d98"} Nov 24 17:24:50 crc kubenswrapper[4777]: I1124 17:24:50.596376 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9412f05c-2d23-4124-abde-2f573bb6855d","Type":"ContainerDied","Data":"f5bc687fc5668ac4935952f9e02557d75c41fa5a98550fa1786c9d926584e7ae"} Nov 24 17:24:50 crc kubenswrapper[4777]: I1124 17:24:50.597998 4777 generic.go:334] "Generic (PLEG): container finished" podID="e36159e1-7be5-49a7-8417-57253b4e7be0" containerID="ab7e9e98a34b1988684706f4ddeaf816d942278c54ee4a8d6a45503c68ca6d2c" exitCode=0 Nov 24 17:24:50 crc kubenswrapper[4777]: I1124 17:24:50.598089 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h2fbd" event={"ID":"e36159e1-7be5-49a7-8417-57253b4e7be0","Type":"ContainerDied","Data":"ab7e9e98a34b1988684706f4ddeaf816d942278c54ee4a8d6a45503c68ca6d2c"} Nov 24 17:24:50 crc kubenswrapper[4777]: I1124 17:24:50.810827 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:24:51 crc kubenswrapper[4777]: I1124 17:24:51.610760 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h2fbd" event={"ID":"e36159e1-7be5-49a7-8417-57253b4e7be0","Type":"ContainerStarted","Data":"98605f3510bf5886f3d7ec67fb9f80aea188b7a6172441ca71246375c34ba891"} Nov 24 17:24:52 crc kubenswrapper[4777]: E1124 17:24:52.661871 4777 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e74119f_b07b_4531_b5fa_5a92a9c6625b.slice/crio-828c4ec9b88e85872599f26f0bc8dbf67daad7b0252833a2b91593572ab5062b.scope\": RecentStats: unable to find data in memory cache]" Nov 24 17:24:52 crc kubenswrapper[4777]: I1124 17:24:52.666205 4777 generic.go:334] "Generic (PLEG): container finished" podID="5e74119f-b07b-4531-b5fa-5a92a9c6625b" containerID="828c4ec9b88e85872599f26f0bc8dbf67daad7b0252833a2b91593572ab5062b" exitCode=0 Nov 24 17:24:52 crc kubenswrapper[4777]: I1124 17:24:52.667025 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-9d788" event={"ID":"5e74119f-b07b-4531-b5fa-5a92a9c6625b","Type":"ContainerDied","Data":"828c4ec9b88e85872599f26f0bc8dbf67daad7b0252833a2b91593572ab5062b"} Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.680014 4777 generic.go:334] "Generic (PLEG): container finished" podID="9412f05c-2d23-4124-abde-2f573bb6855d" containerID="117c081d4641d4be6b7400e1d344b0278b56d84856124fa09bc36ccec9a2698e" exitCode=0 Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.680070 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9412f05c-2d23-4124-abde-2f573bb6855d","Type":"ContainerDied","Data":"117c081d4641d4be6b7400e1d344b0278b56d84856124fa09bc36ccec9a2698e"} Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.786849 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.936232 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9412f05c-2d23-4124-abde-2f573bb6855d-log-httpd\") pod \"9412f05c-2d23-4124-abde-2f573bb6855d\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.936343 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnd4b\" (UniqueName: \"kubernetes.io/projected/9412f05c-2d23-4124-abde-2f573bb6855d-kube-api-access-xnd4b\") pod \"9412f05c-2d23-4124-abde-2f573bb6855d\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.936475 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9412f05c-2d23-4124-abde-2f573bb6855d-run-httpd\") pod \"9412f05c-2d23-4124-abde-2f573bb6855d\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.936542 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-sg-core-conf-yaml\") pod \"9412f05c-2d23-4124-abde-2f573bb6855d\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.936570 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-config-data\") pod \"9412f05c-2d23-4124-abde-2f573bb6855d\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.936608 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-ceilometer-tls-certs\") pod \"9412f05c-2d23-4124-abde-2f573bb6855d\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.936662 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-combined-ca-bundle\") pod \"9412f05c-2d23-4124-abde-2f573bb6855d\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.936766 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-scripts\") pod \"9412f05c-2d23-4124-abde-2f573bb6855d\" (UID: \"9412f05c-2d23-4124-abde-2f573bb6855d\") " Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.937711 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9412f05c-2d23-4124-abde-2f573bb6855d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9412f05c-2d23-4124-abde-2f573bb6855d" (UID: "9412f05c-2d23-4124-abde-2f573bb6855d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.938600 4777 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9412f05c-2d23-4124-abde-2f573bb6855d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.938773 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9412f05c-2d23-4124-abde-2f573bb6855d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9412f05c-2d23-4124-abde-2f573bb6855d" (UID: "9412f05c-2d23-4124-abde-2f573bb6855d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.946086 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-scripts" (OuterVolumeSpecName: "scripts") pod "9412f05c-2d23-4124-abde-2f573bb6855d" (UID: "9412f05c-2d23-4124-abde-2f573bb6855d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.976332 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9412f05c-2d23-4124-abde-2f573bb6855d" (UID: "9412f05c-2d23-4124-abde-2f573bb6855d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:53 crc kubenswrapper[4777]: I1124 17:24:53.977022 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9412f05c-2d23-4124-abde-2f573bb6855d-kube-api-access-xnd4b" (OuterVolumeSpecName: "kube-api-access-xnd4b") pod "9412f05c-2d23-4124-abde-2f573bb6855d" (UID: "9412f05c-2d23-4124-abde-2f573bb6855d"). InnerVolumeSpecName "kube-api-access-xnd4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.042200 4777 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9412f05c-2d23-4124-abde-2f573bb6855d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.042238 4777 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.042249 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.042258 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnd4b\" (UniqueName: \"kubernetes.io/projected/9412f05c-2d23-4124-abde-2f573bb6855d-kube-api-access-xnd4b\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.043070 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="4c707277-46e4-42ea-9076-566b2d01f544" containerName="rabbitmq" containerID="cri-o://e5a70124332b963c74c5101f8eebeb8e795be45b3dcb774cd21bae785822c37b" gracePeriod=604796 Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.048395 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "9412f05c-2d23-4124-abde-2f573bb6855d" (UID: "9412f05c-2d23-4124-abde-2f573bb6855d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.093219 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-config-data" (OuterVolumeSpecName: "config-data") pod "9412f05c-2d23-4124-abde-2f573bb6855d" (UID: "9412f05c-2d23-4124-abde-2f573bb6855d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.112151 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9412f05c-2d23-4124-abde-2f573bb6855d" (UID: "9412f05c-2d23-4124-abde-2f573bb6855d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.118341 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.143098 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt5s2\" (UniqueName: \"kubernetes.io/projected/5e74119f-b07b-4531-b5fa-5a92a9c6625b-kube-api-access-rt5s2\") pod \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.143422 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-scripts\") pod \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.143575 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-config-data\") pod \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.143710 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-combined-ca-bundle\") pod \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.143861 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/5e74119f-b07b-4531-b5fa-5a92a9c6625b-certs\") pod \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\" (UID: \"5e74119f-b07b-4531-b5fa-5a92a9c6625b\") " Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.150273 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.150417 4777 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.150474 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9412f05c-2d23-4124-abde-2f573bb6855d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.156873 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-scripts" (OuterVolumeSpecName: "scripts") pod "5e74119f-b07b-4531-b5fa-5a92a9c6625b" (UID: "5e74119f-b07b-4531-b5fa-5a92a9c6625b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.157020 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e74119f-b07b-4531-b5fa-5a92a9c6625b-kube-api-access-rt5s2" (OuterVolumeSpecName: "kube-api-access-rt5s2") pod "5e74119f-b07b-4531-b5fa-5a92a9c6625b" (UID: "5e74119f-b07b-4531-b5fa-5a92a9c6625b"). InnerVolumeSpecName "kube-api-access-rt5s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.159708 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e74119f-b07b-4531-b5fa-5a92a9c6625b-certs" (OuterVolumeSpecName: "certs") pod "5e74119f-b07b-4531-b5fa-5a92a9c6625b" (UID: "5e74119f-b07b-4531-b5fa-5a92a9c6625b"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.191480 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-config-data" (OuterVolumeSpecName: "config-data") pod "5e74119f-b07b-4531-b5fa-5a92a9c6625b" (UID: "5e74119f-b07b-4531-b5fa-5a92a9c6625b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.207117 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e74119f-b07b-4531-b5fa-5a92a9c6625b" (UID: "5e74119f-b07b-4531-b5fa-5a92a9c6625b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.251739 4777 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/5e74119f-b07b-4531-b5fa-5a92a9c6625b-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.251769 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt5s2\" (UniqueName: \"kubernetes.io/projected/5e74119f-b07b-4531-b5fa-5a92a9c6625b-kube-api-access-rt5s2\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.251779 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.251788 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.251800 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e74119f-b07b-4531-b5fa-5a92a9c6625b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.690451 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-9d788" event={"ID":"5e74119f-b07b-4531-b5fa-5a92a9c6625b","Type":"ContainerDied","Data":"df435a6c8e65906c43b3e360cc272386ddd688ae3c52bc6c8ceb68fadd335047"} Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.690485 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df435a6c8e65906c43b3e360cc272386ddd688ae3c52bc6c8ceb68fadd335047" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.690517 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-9d788" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.694223 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9412f05c-2d23-4124-abde-2f573bb6855d","Type":"ContainerDied","Data":"219de3d81339f912e2e280a46dd2e93b9a6f59dc339241204d410f575d9e5330"} Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.694263 4777 scope.go:117] "RemoveContainer" containerID="a16055db75b25076a62b6f358bf8c8eed9750e38d0a8658f4ebbe207795f1bb8" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.694295 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.722785 4777 scope.go:117] "RemoveContainer" containerID="1a8c50da8aeb8068e5d4150d334c3b86a8ba7c7a17681cd61776dfe674563d98" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.738018 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.747106 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.784693 4777 scope.go:117] "RemoveContainer" containerID="117c081d4641d4be6b7400e1d344b0278b56d84856124fa09bc36ccec9a2698e" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.813595 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:24:54 crc kubenswrapper[4777]: E1124 17:24:54.814855 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e74119f-b07b-4531-b5fa-5a92a9c6625b" containerName="cloudkitty-db-sync" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.814893 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e74119f-b07b-4531-b5fa-5a92a9c6625b" containerName="cloudkitty-db-sync" Nov 24 17:24:54 crc kubenswrapper[4777]: E1124 17:24:54.814934 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="proxy-httpd" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.814947 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="proxy-httpd" Nov 24 17:24:54 crc kubenswrapper[4777]: E1124 17:24:54.815044 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="sg-core" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.815058 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="sg-core" Nov 24 17:24:54 crc kubenswrapper[4777]: E1124 17:24:54.815136 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="ceilometer-notification-agent" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.815149 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="ceilometer-notification-agent" Nov 24 17:24:54 crc kubenswrapper[4777]: E1124 17:24:54.815171 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="ceilometer-central-agent" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.815185 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="ceilometer-central-agent" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.815848 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="ceilometer-central-agent" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.815889 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e74119f-b07b-4531-b5fa-5a92a9c6625b" containerName="cloudkitty-db-sync" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.815917 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="sg-core" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.815959 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="ceilometer-notification-agent" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.816017 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" containerName="proxy-httpd" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.818564 4777 scope.go:117] "RemoveContainer" containerID="f5bc687fc5668ac4935952f9e02557d75c41fa5a98550fa1786c9d926584e7ae" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.827285 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.827854 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.830342 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.830639 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.830946 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.885625 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-storageinit-mqq7w"] Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.894368 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-storageinit-mqq7w"] Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.983398 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-storageinit-fgmx8"] Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.985622 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.987700 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 17:24:54 crc kubenswrapper[4777]: I1124 17:24:54.994550 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-fgmx8"] Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.002050 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg8gm\" (UniqueName: \"kubernetes.io/projected/6f80e34e-82d8-4d61-91b9-323898efdabf-kube-api-access-cg8gm\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.002111 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f80e34e-82d8-4d61-91b9-323898efdabf-scripts\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.002170 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f80e34e-82d8-4d61-91b9-323898efdabf-run-httpd\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.002194 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f80e34e-82d8-4d61-91b9-323898efdabf-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.002274 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f80e34e-82d8-4d61-91b9-323898efdabf-log-httpd\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.002336 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f80e34e-82d8-4d61-91b9-323898efdabf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.002397 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f80e34e-82d8-4d61-91b9-323898efdabf-config-data\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.002480 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6f80e34e-82d8-4d61-91b9-323898efdabf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.105265 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f80e34e-82d8-4d61-91b9-323898efdabf-run-httpd\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.105307 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f80e34e-82d8-4d61-91b9-323898efdabf-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.105380 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f80e34e-82d8-4d61-91b9-323898efdabf-log-httpd\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.105435 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f80e34e-82d8-4d61-91b9-323898efdabf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.105459 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-config-data\") pod \"cloudkitty-storageinit-fgmx8\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.105496 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65rhn\" (UniqueName: \"kubernetes.io/projected/0e5ed725-352f-46ae-83ea-e3b645936482-kube-api-access-65rhn\") pod \"cloudkitty-storageinit-fgmx8\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.105519 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-combined-ca-bundle\") pod \"cloudkitty-storageinit-fgmx8\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.105535 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f80e34e-82d8-4d61-91b9-323898efdabf-config-data\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.105584 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6f80e34e-82d8-4d61-91b9-323898efdabf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.105628 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-scripts\") pod \"cloudkitty-storageinit-fgmx8\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.105646 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg8gm\" (UniqueName: \"kubernetes.io/projected/6f80e34e-82d8-4d61-91b9-323898efdabf-kube-api-access-cg8gm\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.105670 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f80e34e-82d8-4d61-91b9-323898efdabf-scripts\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.105691 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/0e5ed725-352f-46ae-83ea-e3b645936482-certs\") pod \"cloudkitty-storageinit-fgmx8\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.105734 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f80e34e-82d8-4d61-91b9-323898efdabf-run-httpd\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.106624 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f80e34e-82d8-4d61-91b9-323898efdabf-log-httpd\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.113568 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f80e34e-82d8-4d61-91b9-323898efdabf-config-data\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.113581 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f80e34e-82d8-4d61-91b9-323898efdabf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.114631 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6f80e34e-82d8-4d61-91b9-323898efdabf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.115099 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f80e34e-82d8-4d61-91b9-323898efdabf-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.115331 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f80e34e-82d8-4d61-91b9-323898efdabf-scripts\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.125104 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg8gm\" (UniqueName: \"kubernetes.io/projected/6f80e34e-82d8-4d61-91b9-323898efdabf-kube-api-access-cg8gm\") pod \"ceilometer-0\" (UID: \"6f80e34e-82d8-4d61-91b9-323898efdabf\") " pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.158319 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.208184 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-config-data\") pod \"cloudkitty-storageinit-fgmx8\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.208250 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65rhn\" (UniqueName: \"kubernetes.io/projected/0e5ed725-352f-46ae-83ea-e3b645936482-kube-api-access-65rhn\") pod \"cloudkitty-storageinit-fgmx8\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.208277 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-combined-ca-bundle\") pod \"cloudkitty-storageinit-fgmx8\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.208353 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-scripts\") pod \"cloudkitty-storageinit-fgmx8\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.208388 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/0e5ed725-352f-46ae-83ea-e3b645936482-certs\") pod \"cloudkitty-storageinit-fgmx8\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.213684 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-config-data\") pod \"cloudkitty-storageinit-fgmx8\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.214070 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-combined-ca-bundle\") pod \"cloudkitty-storageinit-fgmx8\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.214277 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/0e5ed725-352f-46ae-83ea-e3b645936482-certs\") pod \"cloudkitty-storageinit-fgmx8\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.216396 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-scripts\") pod \"cloudkitty-storageinit-fgmx8\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.249907 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65rhn\" (UniqueName: \"kubernetes.io/projected/0e5ed725-352f-46ae-83ea-e3b645936482-kube-api-access-65rhn\") pod \"cloudkitty-storageinit-fgmx8\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.261820 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="085c54f5-afa5-431a-84d1-75dad9050d2d" path="/var/lib/kubelet/pods/085c54f5-afa5-431a-84d1-75dad9050d2d/volumes" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.262822 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9412f05c-2d23-4124-abde-2f573bb6855d" path="/var/lib/kubelet/pods/9412f05c-2d23-4124-abde-2f573bb6855d/volumes" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.310397 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.706556 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.714153 4777 generic.go:334] "Generic (PLEG): container finished" podID="e36159e1-7be5-49a7-8417-57253b4e7be0" containerID="98605f3510bf5886f3d7ec67fb9f80aea188b7a6172441ca71246375c34ba891" exitCode=0 Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.714245 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h2fbd" event={"ID":"e36159e1-7be5-49a7-8417-57253b4e7be0","Type":"ContainerDied","Data":"98605f3510bf5886f3d7ec67fb9f80aea188b7a6172441ca71246375c34ba891"} Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.729907 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f80e34e-82d8-4d61-91b9-323898efdabf","Type":"ContainerStarted","Data":"aca120af908d28023da9d2477f3828088a2890478d8a1677997550ef6aa367ff"} Nov 24 17:24:55 crc kubenswrapper[4777]: I1124 17:24:55.865114 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-fgmx8"] Nov 24 17:24:56 crc kubenswrapper[4777]: I1124 17:24:56.137978 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="7163366e-9a39-4379-af0a-d6c7c998ab7f" containerName="rabbitmq" containerID="cri-o://de911e01738284bd59a6b0afed7357ce4a53fef2f11ed85c0d07b82a41044436" gracePeriod=604795 Nov 24 17:24:56 crc kubenswrapper[4777]: I1124 17:24:56.746439 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h2fbd" event={"ID":"e36159e1-7be5-49a7-8417-57253b4e7be0","Type":"ContainerStarted","Data":"5d33d49edaff9e300c7160061f0be9fba9b30177e288a940b0e597f28f13ad06"} Nov 24 17:24:56 crc kubenswrapper[4777]: I1124 17:24:56.754899 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-fgmx8" event={"ID":"0e5ed725-352f-46ae-83ea-e3b645936482","Type":"ContainerStarted","Data":"af536934b889d96923c0857686c62be0501f1bd4a996dbc0d04dc714a2a1bdbf"} Nov 24 17:24:56 crc kubenswrapper[4777]: I1124 17:24:56.754939 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-fgmx8" event={"ID":"0e5ed725-352f-46ae-83ea-e3b645936482","Type":"ContainerStarted","Data":"743e688424c3dd794f9a36cb85c1bd4c7d63158493231b4964a7d0e6c7bbfa7c"} Nov 24 17:24:56 crc kubenswrapper[4777]: I1124 17:24:56.773985 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h2fbd" podStartSLOduration=3.111366845 podStartE2EDuration="9.773946623s" podCreationTimestamp="2025-11-24 17:24:47 +0000 UTC" firstStartedPulling="2025-11-24 17:24:49.581547181 +0000 UTC m=+1447.740742230" lastFinishedPulling="2025-11-24 17:24:56.244126949 +0000 UTC m=+1454.403322008" observedRunningTime="2025-11-24 17:24:56.770590626 +0000 UTC m=+1454.929785685" watchObservedRunningTime="2025-11-24 17:24:56.773946623 +0000 UTC m=+1454.933141672" Nov 24 17:24:56 crc kubenswrapper[4777]: I1124 17:24:56.792679 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-storageinit-fgmx8" podStartSLOduration=2.792663244 podStartE2EDuration="2.792663244s" podCreationTimestamp="2025-11-24 17:24:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:24:56.785192818 +0000 UTC m=+1454.944387867" watchObservedRunningTime="2025-11-24 17:24:56.792663244 +0000 UTC m=+1454.951858293" Nov 24 17:24:58 crc kubenswrapper[4777]: I1124 17:24:58.152307 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:24:58 crc kubenswrapper[4777]: I1124 17:24:58.152630 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:24:58 crc kubenswrapper[4777]: I1124 17:24:58.792217 4777 generic.go:334] "Generic (PLEG): container finished" podID="0e5ed725-352f-46ae-83ea-e3b645936482" containerID="af536934b889d96923c0857686c62be0501f1bd4a996dbc0d04dc714a2a1bdbf" exitCode=0 Nov 24 17:24:58 crc kubenswrapper[4777]: I1124 17:24:58.792269 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-fgmx8" event={"ID":"0e5ed725-352f-46ae-83ea-e3b645936482","Type":"ContainerDied","Data":"af536934b889d96923c0857686c62be0501f1bd4a996dbc0d04dc714a2a1bdbf"} Nov 24 17:24:59 crc kubenswrapper[4777]: I1124 17:24:59.209466 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h2fbd" podUID="e36159e1-7be5-49a7-8417-57253b4e7be0" containerName="registry-server" probeResult="failure" output=< Nov 24 17:24:59 crc kubenswrapper[4777]: timeout: failed to connect service ":50051" within 1s Nov 24 17:24:59 crc kubenswrapper[4777]: > Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.351061 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.524919 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-config-data\") pod \"0e5ed725-352f-46ae-83ea-e3b645936482\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.525136 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-scripts\") pod \"0e5ed725-352f-46ae-83ea-e3b645936482\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.525200 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-combined-ca-bundle\") pod \"0e5ed725-352f-46ae-83ea-e3b645936482\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.525247 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65rhn\" (UniqueName: \"kubernetes.io/projected/0e5ed725-352f-46ae-83ea-e3b645936482-kube-api-access-65rhn\") pod \"0e5ed725-352f-46ae-83ea-e3b645936482\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.525293 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/0e5ed725-352f-46ae-83ea-e3b645936482-certs\") pod \"0e5ed725-352f-46ae-83ea-e3b645936482\" (UID: \"0e5ed725-352f-46ae-83ea-e3b645936482\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.530665 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e5ed725-352f-46ae-83ea-e3b645936482-certs" (OuterVolumeSpecName: "certs") pod "0e5ed725-352f-46ae-83ea-e3b645936482" (UID: "0e5ed725-352f-46ae-83ea-e3b645936482"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.533875 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-scripts" (OuterVolumeSpecName: "scripts") pod "0e5ed725-352f-46ae-83ea-e3b645936482" (UID: "0e5ed725-352f-46ae-83ea-e3b645936482"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.534645 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e5ed725-352f-46ae-83ea-e3b645936482-kube-api-access-65rhn" (OuterVolumeSpecName: "kube-api-access-65rhn") pod "0e5ed725-352f-46ae-83ea-e3b645936482" (UID: "0e5ed725-352f-46ae-83ea-e3b645936482"). InnerVolumeSpecName "kube-api-access-65rhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.564351 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e5ed725-352f-46ae-83ea-e3b645936482" (UID: "0e5ed725-352f-46ae-83ea-e3b645936482"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.577051 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-config-data" (OuterVolumeSpecName: "config-data") pod "0e5ed725-352f-46ae-83ea-e3b645936482" (UID: "0e5ed725-352f-46ae-83ea-e3b645936482"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.627681 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.627733 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.627741 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e5ed725-352f-46ae-83ea-e3b645936482-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.627754 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65rhn\" (UniqueName: \"kubernetes.io/projected/0e5ed725-352f-46ae-83ea-e3b645936482-kube-api-access-65rhn\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.627763 4777 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/0e5ed725-352f-46ae-83ea-e3b645936482-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.648083 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.839717 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-config-data\") pod \"4c707277-46e4-42ea-9076-566b2d01f544\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.839998 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-confd\") pod \"4c707277-46e4-42ea-9076-566b2d01f544\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.840099 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4c707277-46e4-42ea-9076-566b2d01f544-erlang-cookie-secret\") pod \"4c707277-46e4-42ea-9076-566b2d01f544\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.840123 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-server-conf\") pod \"4c707277-46e4-42ea-9076-566b2d01f544\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.840159 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-erlang-cookie\") pod \"4c707277-46e4-42ea-9076-566b2d01f544\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.840186 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ltf9\" (UniqueName: \"kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-kube-api-access-2ltf9\") pod \"4c707277-46e4-42ea-9076-566b2d01f544\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.840281 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-tls\") pod \"4c707277-46e4-42ea-9076-566b2d01f544\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.840313 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4c707277-46e4-42ea-9076-566b2d01f544-pod-info\") pod \"4c707277-46e4-42ea-9076-566b2d01f544\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.840330 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-plugins-conf\") pod \"4c707277-46e4-42ea-9076-566b2d01f544\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.841069 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\") pod \"4c707277-46e4-42ea-9076-566b2d01f544\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.841112 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-plugins\") pod \"4c707277-46e4-42ea-9076-566b2d01f544\" (UID: \"4c707277-46e4-42ea-9076-566b2d01f544\") " Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.843915 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "4c707277-46e4-42ea-9076-566b2d01f544" (UID: "4c707277-46e4-42ea-9076-566b2d01f544"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.845747 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "4c707277-46e4-42ea-9076-566b2d01f544" (UID: "4c707277-46e4-42ea-9076-566b2d01f544"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.857842 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-kube-api-access-2ltf9" (OuterVolumeSpecName: "kube-api-access-2ltf9") pod "4c707277-46e4-42ea-9076-566b2d01f544" (UID: "4c707277-46e4-42ea-9076-566b2d01f544"). InnerVolumeSpecName "kube-api-access-2ltf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.858896 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/4c707277-46e4-42ea-9076-566b2d01f544-pod-info" (OuterVolumeSpecName: "pod-info") pod "4c707277-46e4-42ea-9076-566b2d01f544" (UID: "4c707277-46e4-42ea-9076-566b2d01f544"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.866016 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "4c707277-46e4-42ea-9076-566b2d01f544" (UID: "4c707277-46e4-42ea-9076-566b2d01f544"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.874197 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "4c707277-46e4-42ea-9076-566b2d01f544" (UID: "4c707277-46e4-42ea-9076-566b2d01f544"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.878235 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c707277-46e4-42ea-9076-566b2d01f544-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "4c707277-46e4-42ea-9076-566b2d01f544" (UID: "4c707277-46e4-42ea-9076-566b2d01f544"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.879252 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-fgmx8" event={"ID":"0e5ed725-352f-46ae-83ea-e3b645936482","Type":"ContainerDied","Data":"743e688424c3dd794f9a36cb85c1bd4c7d63158493231b4964a7d0e6c7bbfa7c"} Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.880081 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="743e688424c3dd794f9a36cb85c1bd4c7d63158493231b4964a7d0e6c7bbfa7c" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.880131 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-fgmx8" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.890596 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f80e34e-82d8-4d61-91b9-323898efdabf","Type":"ContainerStarted","Data":"6cba41277a777bf13ddf70eb338018794274eb3d4bf89910c6801558d9e0179c"} Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.890649 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f80e34e-82d8-4d61-91b9-323898efdabf","Type":"ContainerStarted","Data":"aa13fb3e79f5a07ac0554b03bc884d760b3d710721c5adc7eda14ec2956427af"} Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.898485 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.898386 4777 generic.go:334] "Generic (PLEG): container finished" podID="4c707277-46e4-42ea-9076-566b2d01f544" containerID="e5a70124332b963c74c5101f8eebeb8e795be45b3dcb774cd21bae785822c37b" exitCode=0 Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.898596 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4c707277-46e4-42ea-9076-566b2d01f544","Type":"ContainerDied","Data":"e5a70124332b963c74c5101f8eebeb8e795be45b3dcb774cd21bae785822c37b"} Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.898623 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4c707277-46e4-42ea-9076-566b2d01f544","Type":"ContainerDied","Data":"d8494ef2a24b9eec544579065892957d1c4431eea4c43c065715a0430702f202"} Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.898640 4777 scope.go:117] "RemoveContainer" containerID="e5a70124332b963c74c5101f8eebeb8e795be45b3dcb774cd21bae785822c37b" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.938792 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-config-data" (OuterVolumeSpecName: "config-data") pod "4c707277-46e4-42ea-9076-566b2d01f544" (UID: "4c707277-46e4-42ea-9076-566b2d01f544"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.948869 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.948899 4777 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4c707277-46e4-42ea-9076-566b2d01f544-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.948910 4777 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.948919 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ltf9\" (UniqueName: \"kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-kube-api-access-2ltf9\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.948928 4777 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.948936 4777 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4c707277-46e4-42ea-9076-566b2d01f544-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.948943 4777 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.948951 4777 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.962446 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7" (OuterVolumeSpecName: "persistence") pod "4c707277-46e4-42ea-9076-566b2d01f544" (UID: "4c707277-46e4-42ea-9076-566b2d01f544"). InnerVolumeSpecName "pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.975074 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-server-conf" (OuterVolumeSpecName: "server-conf") pod "4c707277-46e4-42ea-9076-566b2d01f544" (UID: "4c707277-46e4-42ea-9076-566b2d01f544"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.994828 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 17:25:00 crc kubenswrapper[4777]: I1124 17:25:00.995097 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="2aa8a135-f57b-4d6b-9444-4fbdfc542476" containerName="cloudkitty-proc" containerID="cri-o://896237ee47f1b5107998401ff73591a93018704bd939c2fb9289492ce4fbcd36" gracePeriod=30 Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.007299 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.007532 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="94ad549a-1f18-4ad2-8d56-844e83d35aec" containerName="cloudkitty-api-log" containerID="cri-o://1280b04734fef908aac2b889626d0963282ca59afe1eb90ea1366d3c28b44536" gracePeriod=30 Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.008019 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="94ad549a-1f18-4ad2-8d56-844e83d35aec" containerName="cloudkitty-api" containerID="cri-o://167e60e62b59ed489e4a5c43f0ae11f7a45c8d96c0ba0017e044a46720f1c382" gracePeriod=30 Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.051269 4777 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\") on node \"crc\" " Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.053246 4777 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4c707277-46e4-42ea-9076-566b2d01f544-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.078518 4777 scope.go:117] "RemoveContainer" containerID="9dfcd853b3fc5572da0af858e5836f1849e0716f5fca857126f422e455bcc6b7" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.109151 4777 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.111760 4777 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7") on node "crc" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.137257 4777 scope.go:117] "RemoveContainer" containerID="e5a70124332b963c74c5101f8eebeb8e795be45b3dcb774cd21bae785822c37b" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.144151 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "4c707277-46e4-42ea-9076-566b2d01f544" (UID: "4c707277-46e4-42ea-9076-566b2d01f544"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:01 crc kubenswrapper[4777]: E1124 17:25:01.144277 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5a70124332b963c74c5101f8eebeb8e795be45b3dcb774cd21bae785822c37b\": container with ID starting with e5a70124332b963c74c5101f8eebeb8e795be45b3dcb774cd21bae785822c37b not found: ID does not exist" containerID="e5a70124332b963c74c5101f8eebeb8e795be45b3dcb774cd21bae785822c37b" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.144390 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5a70124332b963c74c5101f8eebeb8e795be45b3dcb774cd21bae785822c37b"} err="failed to get container status \"e5a70124332b963c74c5101f8eebeb8e795be45b3dcb774cd21bae785822c37b\": rpc error: code = NotFound desc = could not find container \"e5a70124332b963c74c5101f8eebeb8e795be45b3dcb774cd21bae785822c37b\": container with ID starting with e5a70124332b963c74c5101f8eebeb8e795be45b3dcb774cd21bae785822c37b not found: ID does not exist" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.144624 4777 scope.go:117] "RemoveContainer" containerID="9dfcd853b3fc5572da0af858e5836f1849e0716f5fca857126f422e455bcc6b7" Nov 24 17:25:01 crc kubenswrapper[4777]: E1124 17:25:01.145902 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dfcd853b3fc5572da0af858e5836f1849e0716f5fca857126f422e455bcc6b7\": container with ID starting with 9dfcd853b3fc5572da0af858e5836f1849e0716f5fca857126f422e455bcc6b7 not found: ID does not exist" containerID="9dfcd853b3fc5572da0af858e5836f1849e0716f5fca857126f422e455bcc6b7" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.145999 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dfcd853b3fc5572da0af858e5836f1849e0716f5fca857126f422e455bcc6b7"} err="failed to get container status \"9dfcd853b3fc5572da0af858e5836f1849e0716f5fca857126f422e455bcc6b7\": rpc error: code = NotFound desc = could not find container \"9dfcd853b3fc5572da0af858e5836f1849e0716f5fca857126f422e455bcc6b7\": container with ID starting with 9dfcd853b3fc5572da0af858e5836f1849e0716f5fca857126f422e455bcc6b7 not found: ID does not exist" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.155579 4777 reconciler_common.go:293] "Volume detached for volume \"pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.155613 4777 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4c707277-46e4-42ea-9076-566b2d01f544-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.286164 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.286209 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.290495 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:25:01 crc kubenswrapper[4777]: E1124 17:25:01.290896 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c707277-46e4-42ea-9076-566b2d01f544" containerName="rabbitmq" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.290912 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c707277-46e4-42ea-9076-566b2d01f544" containerName="rabbitmq" Nov 24 17:25:01 crc kubenswrapper[4777]: E1124 17:25:01.290923 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c707277-46e4-42ea-9076-566b2d01f544" containerName="setup-container" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.290929 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c707277-46e4-42ea-9076-566b2d01f544" containerName="setup-container" Nov 24 17:25:01 crc kubenswrapper[4777]: E1124 17:25:01.291047 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e5ed725-352f-46ae-83ea-e3b645936482" containerName="cloudkitty-storageinit" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.291058 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e5ed725-352f-46ae-83ea-e3b645936482" containerName="cloudkitty-storageinit" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.291233 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e5ed725-352f-46ae-83ea-e3b645936482" containerName="cloudkitty-storageinit" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.291260 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c707277-46e4-42ea-9076-566b2d01f544" containerName="rabbitmq" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.292375 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.297103 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.297366 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.297382 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.297504 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.297629 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-cks9w" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.297521 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.297515 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.318308 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.463765 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78712b89-27c4-4481-b152-320fb7fee9fc-pod-info\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.463831 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/78712b89-27c4-4481-b152-320fb7fee9fc-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.463866 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/78712b89-27c4-4481-b152-320fb7fee9fc-server-conf\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.464111 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs9nh\" (UniqueName: \"kubernetes.io/projected/78712b89-27c4-4481-b152-320fb7fee9fc-kube-api-access-hs9nh\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.464172 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78712b89-27c4-4481-b152-320fb7fee9fc-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.464336 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78712b89-27c4-4481-b152-320fb7fee9fc-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.464386 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78712b89-27c4-4481-b152-320fb7fee9fc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.464446 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78712b89-27c4-4481-b152-320fb7fee9fc-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.464518 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.464691 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78712b89-27c4-4481-b152-320fb7fee9fc-config-data\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.464769 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78712b89-27c4-4481-b152-320fb7fee9fc-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.566921 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs9nh\" (UniqueName: \"kubernetes.io/projected/78712b89-27c4-4481-b152-320fb7fee9fc-kube-api-access-hs9nh\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.567107 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78712b89-27c4-4481-b152-320fb7fee9fc-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.567160 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78712b89-27c4-4481-b152-320fb7fee9fc-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.567184 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78712b89-27c4-4481-b152-320fb7fee9fc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.567221 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78712b89-27c4-4481-b152-320fb7fee9fc-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.567267 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.567393 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78712b89-27c4-4481-b152-320fb7fee9fc-config-data\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.567430 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78712b89-27c4-4481-b152-320fb7fee9fc-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.567501 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78712b89-27c4-4481-b152-320fb7fee9fc-pod-info\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.567533 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/78712b89-27c4-4481-b152-320fb7fee9fc-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.567578 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/78712b89-27c4-4481-b152-320fb7fee9fc-server-conf\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.567747 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78712b89-27c4-4481-b152-320fb7fee9fc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.568265 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78712b89-27c4-4481-b152-320fb7fee9fc-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.569319 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78712b89-27c4-4481-b152-320fb7fee9fc-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.569424 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78712b89-27c4-4481-b152-320fb7fee9fc-config-data\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.570342 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/78712b89-27c4-4481-b152-320fb7fee9fc-server-conf\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.571772 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/78712b89-27c4-4481-b152-320fb7fee9fc-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.572248 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78712b89-27c4-4481-b152-320fb7fee9fc-pod-info\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.573109 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78712b89-27c4-4481-b152-320fb7fee9fc-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.573725 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.573780 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/38e94766a61c3e0d686381901a2b5cf624719d54de394fd2452d21d58a78d20d/globalmount\"" pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.575560 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78712b89-27c4-4481-b152-320fb7fee9fc-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.587468 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs9nh\" (UniqueName: \"kubernetes.io/projected/78712b89-27c4-4481-b152-320fb7fee9fc-kube-api-access-hs9nh\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.605064 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="7163366e-9a39-4379-af0a-d6c7c998ab7f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.113:5671: connect: connection refused" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.618700 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a427a88-40e9-4f6a-a72e-dae67c5440d7\") pod \"rabbitmq-server-0\" (UID: \"78712b89-27c4-4481-b152-320fb7fee9fc\") " pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.648476 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.890413 4777 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-api-0" podUID="94ad549a-1f18-4ad2-8d56-844e83d35aec" containerName="cloudkitty-api" probeResult="failure" output="Get \"https://10.217.0.199:8889/healthcheck\": read tcp 10.217.0.2:50722->10.217.0.199:8889: read: connection reset by peer" Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.927500 4777 generic.go:334] "Generic (PLEG): container finished" podID="2aa8a135-f57b-4d6b-9444-4fbdfc542476" containerID="896237ee47f1b5107998401ff73591a93018704bd939c2fb9289492ce4fbcd36" exitCode=0 Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.927567 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"2aa8a135-f57b-4d6b-9444-4fbdfc542476","Type":"ContainerDied","Data":"896237ee47f1b5107998401ff73591a93018704bd939c2fb9289492ce4fbcd36"} Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.932490 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f80e34e-82d8-4d61-91b9-323898efdabf","Type":"ContainerStarted","Data":"f28a78eeccf1ed5a75c8cab2659c851ab11a201c3a4e3916864a2a5939d31280"} Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.936595 4777 generic.go:334] "Generic (PLEG): container finished" podID="94ad549a-1f18-4ad2-8d56-844e83d35aec" containerID="1280b04734fef908aac2b889626d0963282ca59afe1eb90ea1366d3c28b44536" exitCode=143 Nov 24 17:25:01 crc kubenswrapper[4777]: I1124 17:25:01.936666 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"94ad549a-1f18-4ad2-8d56-844e83d35aec","Type":"ContainerDied","Data":"1280b04734fef908aac2b889626d0963282ca59afe1eb90ea1366d3c28b44536"} Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.133705 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.380346 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.413878 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.492220 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-config-data-custom\") pod \"94ad549a-1f18-4ad2-8d56-844e83d35aec\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.492285 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-scripts\") pod \"94ad549a-1f18-4ad2-8d56-844e83d35aec\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.492333 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vmhp\" (UniqueName: \"kubernetes.io/projected/2aa8a135-f57b-4d6b-9444-4fbdfc542476-kube-api-access-5vmhp\") pod \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.492367 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-internal-tls-certs\") pod \"94ad549a-1f18-4ad2-8d56-844e83d35aec\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.492427 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-scripts\") pod \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.492452 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-public-tls-certs\") pod \"94ad549a-1f18-4ad2-8d56-844e83d35aec\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.492627 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwqlx\" (UniqueName: \"kubernetes.io/projected/94ad549a-1f18-4ad2-8d56-844e83d35aec-kube-api-access-bwqlx\") pod \"94ad549a-1f18-4ad2-8d56-844e83d35aec\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.492655 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-config-data\") pod \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.492700 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-config-data-custom\") pod \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.492721 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/2aa8a135-f57b-4d6b-9444-4fbdfc542476-certs\") pod \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.492736 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-combined-ca-bundle\") pod \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\" (UID: \"2aa8a135-f57b-4d6b-9444-4fbdfc542476\") " Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.492787 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94ad549a-1f18-4ad2-8d56-844e83d35aec-logs\") pod \"94ad549a-1f18-4ad2-8d56-844e83d35aec\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.492804 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/94ad549a-1f18-4ad2-8d56-844e83d35aec-certs\") pod \"94ad549a-1f18-4ad2-8d56-844e83d35aec\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.492945 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-combined-ca-bundle\") pod \"94ad549a-1f18-4ad2-8d56-844e83d35aec\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.493000 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-config-data\") pod \"94ad549a-1f18-4ad2-8d56-844e83d35aec\" (UID: \"94ad549a-1f18-4ad2-8d56-844e83d35aec\") " Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.496701 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aa8a135-f57b-4d6b-9444-4fbdfc542476-kube-api-access-5vmhp" (OuterVolumeSpecName: "kube-api-access-5vmhp") pod "2aa8a135-f57b-4d6b-9444-4fbdfc542476" (UID: "2aa8a135-f57b-4d6b-9444-4fbdfc542476"). InnerVolumeSpecName "kube-api-access-5vmhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.497160 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2aa8a135-f57b-4d6b-9444-4fbdfc542476" (UID: "2aa8a135-f57b-4d6b-9444-4fbdfc542476"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.498471 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "94ad549a-1f18-4ad2-8d56-844e83d35aec" (UID: "94ad549a-1f18-4ad2-8d56-844e83d35aec"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.499020 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94ad549a-1f18-4ad2-8d56-844e83d35aec-logs" (OuterVolumeSpecName: "logs") pod "94ad549a-1f18-4ad2-8d56-844e83d35aec" (UID: "94ad549a-1f18-4ad2-8d56-844e83d35aec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.505146 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aa8a135-f57b-4d6b-9444-4fbdfc542476-certs" (OuterVolumeSpecName: "certs") pod "2aa8a135-f57b-4d6b-9444-4fbdfc542476" (UID: "2aa8a135-f57b-4d6b-9444-4fbdfc542476"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.505499 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94ad549a-1f18-4ad2-8d56-844e83d35aec-kube-api-access-bwqlx" (OuterVolumeSpecName: "kube-api-access-bwqlx") pod "94ad549a-1f18-4ad2-8d56-844e83d35aec" (UID: "94ad549a-1f18-4ad2-8d56-844e83d35aec"). InnerVolumeSpecName "kube-api-access-bwqlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.506191 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94ad549a-1f18-4ad2-8d56-844e83d35aec-certs" (OuterVolumeSpecName: "certs") pod "94ad549a-1f18-4ad2-8d56-844e83d35aec" (UID: "94ad549a-1f18-4ad2-8d56-844e83d35aec"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.518641 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-scripts" (OuterVolumeSpecName: "scripts") pod "94ad549a-1f18-4ad2-8d56-844e83d35aec" (UID: "94ad549a-1f18-4ad2-8d56-844e83d35aec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.518654 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-scripts" (OuterVolumeSpecName: "scripts") pod "2aa8a135-f57b-4d6b-9444-4fbdfc542476" (UID: "2aa8a135-f57b-4d6b-9444-4fbdfc542476"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.538339 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-config-data" (OuterVolumeSpecName: "config-data") pod "2aa8a135-f57b-4d6b-9444-4fbdfc542476" (UID: "2aa8a135-f57b-4d6b-9444-4fbdfc542476"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.538366 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2aa8a135-f57b-4d6b-9444-4fbdfc542476" (UID: "2aa8a135-f57b-4d6b-9444-4fbdfc542476"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.554020 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94ad549a-1f18-4ad2-8d56-844e83d35aec" (UID: "94ad549a-1f18-4ad2-8d56-844e83d35aec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.599075 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.599110 4777 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.599121 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.599131 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vmhp\" (UniqueName: \"kubernetes.io/projected/2aa8a135-f57b-4d6b-9444-4fbdfc542476-kube-api-access-5vmhp\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.599150 4777 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.599159 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwqlx\" (UniqueName: \"kubernetes.io/projected/94ad549a-1f18-4ad2-8d56-844e83d35aec-kube-api-access-bwqlx\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.599167 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.599175 4777 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.599182 4777 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/2aa8a135-f57b-4d6b-9444-4fbdfc542476-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.599190 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aa8a135-f57b-4d6b-9444-4fbdfc542476-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.599200 4777 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94ad549a-1f18-4ad2-8d56-844e83d35aec-logs\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.599208 4777 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/94ad549a-1f18-4ad2-8d56-844e83d35aec-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.778765 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "94ad549a-1f18-4ad2-8d56-844e83d35aec" (UID: "94ad549a-1f18-4ad2-8d56-844e83d35aec"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.788309 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-config-data" (OuterVolumeSpecName: "config-data") pod "94ad549a-1f18-4ad2-8d56-844e83d35aec" (UID: "94ad549a-1f18-4ad2-8d56-844e83d35aec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.803288 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.804608 4777 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.818528 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "94ad549a-1f18-4ad2-8d56-844e83d35aec" (UID: "94ad549a-1f18-4ad2-8d56-844e83d35aec"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.906543 4777 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/94ad549a-1f18-4ad2-8d56-844e83d35aec-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.952684 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"78712b89-27c4-4481-b152-320fb7fee9fc","Type":"ContainerStarted","Data":"c152c7b0d99ccc49d240d6546609bf127e5fa7c6a0d01ca43c4af8e5d0692326"} Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.957658 4777 generic.go:334] "Generic (PLEG): container finished" podID="94ad549a-1f18-4ad2-8d56-844e83d35aec" containerID="167e60e62b59ed489e4a5c43f0ae11f7a45c8d96c0ba0017e044a46720f1c382" exitCode=0 Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.957812 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.957850 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"94ad549a-1f18-4ad2-8d56-844e83d35aec","Type":"ContainerDied","Data":"167e60e62b59ed489e4a5c43f0ae11f7a45c8d96c0ba0017e044a46720f1c382"} Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.958639 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"94ad549a-1f18-4ad2-8d56-844e83d35aec","Type":"ContainerDied","Data":"e053e3cda387b7ba7b7e17346f4ee0b28e592e30d06f6edf3f759e1387a710c3"} Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.958752 4777 scope.go:117] "RemoveContainer" containerID="167e60e62b59ed489e4a5c43f0ae11f7a45c8d96c0ba0017e044a46720f1c382" Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.962426 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"2aa8a135-f57b-4d6b-9444-4fbdfc542476","Type":"ContainerDied","Data":"87cdfb2999f0d2da241c0e6943520c119fe0205538a1047748a3216a36a7ebd0"} Nov 24 17:25:02 crc kubenswrapper[4777]: I1124 17:25:02.962513 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.002099 4777 scope.go:117] "RemoveContainer" containerID="1280b04734fef908aac2b889626d0963282ca59afe1eb90ea1366d3c28b44536" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.049516 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.057239 4777 scope.go:117] "RemoveContainer" containerID="167e60e62b59ed489e4a5c43f0ae11f7a45c8d96c0ba0017e044a46720f1c382" Nov 24 17:25:03 crc kubenswrapper[4777]: E1124 17:25:03.059141 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"167e60e62b59ed489e4a5c43f0ae11f7a45c8d96c0ba0017e044a46720f1c382\": container with ID starting with 167e60e62b59ed489e4a5c43f0ae11f7a45c8d96c0ba0017e044a46720f1c382 not found: ID does not exist" containerID="167e60e62b59ed489e4a5c43f0ae11f7a45c8d96c0ba0017e044a46720f1c382" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.059181 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"167e60e62b59ed489e4a5c43f0ae11f7a45c8d96c0ba0017e044a46720f1c382"} err="failed to get container status \"167e60e62b59ed489e4a5c43f0ae11f7a45c8d96c0ba0017e044a46720f1c382\": rpc error: code = NotFound desc = could not find container \"167e60e62b59ed489e4a5c43f0ae11f7a45c8d96c0ba0017e044a46720f1c382\": container with ID starting with 167e60e62b59ed489e4a5c43f0ae11f7a45c8d96c0ba0017e044a46720f1c382 not found: ID does not exist" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.059216 4777 scope.go:117] "RemoveContainer" containerID="1280b04734fef908aac2b889626d0963282ca59afe1eb90ea1366d3c28b44536" Nov 24 17:25:03 crc kubenswrapper[4777]: E1124 17:25:03.059474 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1280b04734fef908aac2b889626d0963282ca59afe1eb90ea1366d3c28b44536\": container with ID starting with 1280b04734fef908aac2b889626d0963282ca59afe1eb90ea1366d3c28b44536 not found: ID does not exist" containerID="1280b04734fef908aac2b889626d0963282ca59afe1eb90ea1366d3c28b44536" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.060200 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1280b04734fef908aac2b889626d0963282ca59afe1eb90ea1366d3c28b44536"} err="failed to get container status \"1280b04734fef908aac2b889626d0963282ca59afe1eb90ea1366d3c28b44536\": rpc error: code = NotFound desc = could not find container \"1280b04734fef908aac2b889626d0963282ca59afe1eb90ea1366d3c28b44536\": container with ID starting with 1280b04734fef908aac2b889626d0963282ca59afe1eb90ea1366d3c28b44536 not found: ID does not exist" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.060215 4777 scope.go:117] "RemoveContainer" containerID="896237ee47f1b5107998401ff73591a93018704bd939c2fb9289492ce4fbcd36" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.083082 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.099594 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.109203 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 17:25:03 crc kubenswrapper[4777]: E1124 17:25:03.109711 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94ad549a-1f18-4ad2-8d56-844e83d35aec" containerName="cloudkitty-api" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.109734 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="94ad549a-1f18-4ad2-8d56-844e83d35aec" containerName="cloudkitty-api" Nov 24 17:25:03 crc kubenswrapper[4777]: E1124 17:25:03.109757 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aa8a135-f57b-4d6b-9444-4fbdfc542476" containerName="cloudkitty-proc" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.109765 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aa8a135-f57b-4d6b-9444-4fbdfc542476" containerName="cloudkitty-proc" Nov 24 17:25:03 crc kubenswrapper[4777]: E1124 17:25:03.109806 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94ad549a-1f18-4ad2-8d56-844e83d35aec" containerName="cloudkitty-api-log" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.109816 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="94ad549a-1f18-4ad2-8d56-844e83d35aec" containerName="cloudkitty-api-log" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.110061 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aa8a135-f57b-4d6b-9444-4fbdfc542476" containerName="cloudkitty-proc" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.110352 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="94ad549a-1f18-4ad2-8d56-844e83d35aec" containerName="cloudkitty-api-log" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.110374 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="94ad549a-1f18-4ad2-8d56-844e83d35aec" containerName="cloudkitty-api" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.111835 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.115448 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-public-svc" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.115820 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.116236 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-l2r9w" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.116247 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.116886 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-internal-svc" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.120143 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.120306 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.123683 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.138017 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.147144 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.148677 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.150999 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.161497 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.219492 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lrz6\" (UniqueName: \"kubernetes.io/projected/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-kube-api-access-6lrz6\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.219546 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.219585 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-scripts\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.219611 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-logs\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.219631 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.219661 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-certs\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.219715 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-config-data\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.219757 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.219790 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.260306 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aa8a135-f57b-4d6b-9444-4fbdfc542476" path="/var/lib/kubelet/pods/2aa8a135-f57b-4d6b-9444-4fbdfc542476/volumes" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.261386 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c707277-46e4-42ea-9076-566b2d01f544" path="/var/lib/kubelet/pods/4c707277-46e4-42ea-9076-566b2d01f544/volumes" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.262259 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94ad549a-1f18-4ad2-8d56-844e83d35aec" path="/var/lib/kubelet/pods/94ad549a-1f18-4ad2-8d56-844e83d35aec/volumes" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.321228 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-config-data\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.321296 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.321333 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.322260 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c366c57-2d20-4e76-b90f-c90fc01347fc-config-data\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.322289 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55lzq\" (UniqueName: \"kubernetes.io/projected/1c366c57-2d20-4e76-b90f-c90fc01347fc-kube-api-access-55lzq\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.322311 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c366c57-2d20-4e76-b90f-c90fc01347fc-scripts\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.322454 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lrz6\" (UniqueName: \"kubernetes.io/projected/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-kube-api-access-6lrz6\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.322545 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.322595 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/1c366c57-2d20-4e76-b90f-c90fc01347fc-certs\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.322635 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-scripts\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.322690 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-logs\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.322723 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.322771 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-certs\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.322800 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c366c57-2d20-4e76-b90f-c90fc01347fc-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.322822 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c366c57-2d20-4e76-b90f-c90fc01347fc-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.323857 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-logs\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.327125 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-certs\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.327194 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.327220 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.329141 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-config-data\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.329189 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.329293 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.330057 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-scripts\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.424640 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c366c57-2d20-4e76-b90f-c90fc01347fc-config-data\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.424704 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55lzq\" (UniqueName: \"kubernetes.io/projected/1c366c57-2d20-4e76-b90f-c90fc01347fc-kube-api-access-55lzq\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.424726 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c366c57-2d20-4e76-b90f-c90fc01347fc-scripts\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.424802 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/1c366c57-2d20-4e76-b90f-c90fc01347fc-certs\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.424943 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c366c57-2d20-4e76-b90f-c90fc01347fc-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.424984 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c366c57-2d20-4e76-b90f-c90fc01347fc-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.449019 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c366c57-2d20-4e76-b90f-c90fc01347fc-config-data\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.449332 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c366c57-2d20-4e76-b90f-c90fc01347fc-scripts\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.450203 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/1c366c57-2d20-4e76-b90f-c90fc01347fc-certs\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.451548 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lrz6\" (UniqueName: \"kubernetes.io/projected/5f4b4358-805a-4a55-a5a8-9200c0c5e5b5-kube-api-access-6lrz6\") pod \"cloudkitty-api-0\" (UID: \"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5\") " pod="openstack/cloudkitty-api-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.452950 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c366c57-2d20-4e76-b90f-c90fc01347fc-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.453764 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55lzq\" (UniqueName: \"kubernetes.io/projected/1c366c57-2d20-4e76-b90f-c90fc01347fc-kube-api-access-55lzq\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.453947 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c366c57-2d20-4e76-b90f-c90fc01347fc-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"1c366c57-2d20-4e76-b90f-c90fc01347fc\") " pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.471649 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Nov 24 17:25:03 crc kubenswrapper[4777]: I1124 17:25:03.734538 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.039652 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f80e34e-82d8-4d61-91b9-323898efdabf","Type":"ContainerStarted","Data":"65b3834fa87348dd122fcbbeab43c76a06c4655b750bcecee96258dbdf4cbf78"} Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.041062 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.047083 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"78712b89-27c4-4481-b152-320fb7fee9fc","Type":"ContainerStarted","Data":"bf8d665b16888241cb30e97b5e1fe0a688ac6405fe01012e413ef52b2e3c76cb"} Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.053107 4777 generic.go:334] "Generic (PLEG): container finished" podID="7163366e-9a39-4379-af0a-d6c7c998ab7f" containerID="de911e01738284bd59a6b0afed7357ce4a53fef2f11ed85c0d07b82a41044436" exitCode=0 Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.053170 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7163366e-9a39-4379-af0a-d6c7c998ab7f","Type":"ContainerDied","Data":"de911e01738284bd59a6b0afed7357ce4a53fef2f11ed85c0d07b82a41044436"} Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.087978 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.289871885 podStartE2EDuration="10.087946294s" podCreationTimestamp="2025-11-24 17:24:54 +0000 UTC" firstStartedPulling="2025-11-24 17:24:55.695402859 +0000 UTC m=+1453.854597908" lastFinishedPulling="2025-11-24 17:25:02.493477278 +0000 UTC m=+1460.652672317" observedRunningTime="2025-11-24 17:25:04.077219673 +0000 UTC m=+1462.236414742" watchObservedRunningTime="2025-11-24 17:25:04.087946294 +0000 UTC m=+1462.247141343" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.184595 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.193305 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.347367 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-config-data\") pod \"7163366e-9a39-4379-af0a-d6c7c998ab7f\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.347454 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-plugins-conf\") pod \"7163366e-9a39-4379-af0a-d6c7c998ab7f\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.348672 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "7163366e-9a39-4379-af0a-d6c7c998ab7f" (UID: "7163366e-9a39-4379-af0a-d6c7c998ab7f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.348856 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\") pod \"7163366e-9a39-4379-af0a-d6c7c998ab7f\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.349236 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbf6t\" (UniqueName: \"kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-kube-api-access-cbf6t\") pod \"7163366e-9a39-4379-af0a-d6c7c998ab7f\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.349294 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7163366e-9a39-4379-af0a-d6c7c998ab7f-erlang-cookie-secret\") pod \"7163366e-9a39-4379-af0a-d6c7c998ab7f\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.349321 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-erlang-cookie\") pod \"7163366e-9a39-4379-af0a-d6c7c998ab7f\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.349444 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-tls\") pod \"7163366e-9a39-4379-af0a-d6c7c998ab7f\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.349476 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7163366e-9a39-4379-af0a-d6c7c998ab7f-pod-info\") pod \"7163366e-9a39-4379-af0a-d6c7c998ab7f\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.349557 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-confd\") pod \"7163366e-9a39-4379-af0a-d6c7c998ab7f\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.349594 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-server-conf\") pod \"7163366e-9a39-4379-af0a-d6c7c998ab7f\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.349660 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-plugins\") pod \"7163366e-9a39-4379-af0a-d6c7c998ab7f\" (UID: \"7163366e-9a39-4379-af0a-d6c7c998ab7f\") " Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.350313 4777 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.352683 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "7163366e-9a39-4379-af0a-d6c7c998ab7f" (UID: "7163366e-9a39-4379-af0a-d6c7c998ab7f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.357827 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/7163366e-9a39-4379-af0a-d6c7c998ab7f-pod-info" (OuterVolumeSpecName: "pod-info") pod "7163366e-9a39-4379-af0a-d6c7c998ab7f" (UID: "7163366e-9a39-4379-af0a-d6c7c998ab7f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.358868 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "7163366e-9a39-4379-af0a-d6c7c998ab7f" (UID: "7163366e-9a39-4379-af0a-d6c7c998ab7f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.359348 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7163366e-9a39-4379-af0a-d6c7c998ab7f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "7163366e-9a39-4379-af0a-d6c7c998ab7f" (UID: "7163366e-9a39-4379-af0a-d6c7c998ab7f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.360583 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-kube-api-access-cbf6t" (OuterVolumeSpecName: "kube-api-access-cbf6t") pod "7163366e-9a39-4379-af0a-d6c7c998ab7f" (UID: "7163366e-9a39-4379-af0a-d6c7c998ab7f"). InnerVolumeSpecName "kube-api-access-cbf6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.362264 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "7163366e-9a39-4379-af0a-d6c7c998ab7f" (UID: "7163366e-9a39-4379-af0a-d6c7c998ab7f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.375956 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f" (OuterVolumeSpecName: "persistence") pod "7163366e-9a39-4379-af0a-d6c7c998ab7f" (UID: "7163366e-9a39-4379-af0a-d6c7c998ab7f"). InnerVolumeSpecName "pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.386860 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-config-data" (OuterVolumeSpecName: "config-data") pod "7163366e-9a39-4379-af0a-d6c7c998ab7f" (UID: "7163366e-9a39-4379-af0a-d6c7c998ab7f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.418582 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-server-conf" (OuterVolumeSpecName: "server-conf") pod "7163366e-9a39-4379-af0a-d6c7c998ab7f" (UID: "7163366e-9a39-4379-af0a-d6c7c998ab7f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.418956 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.452645 4777 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\") on node \"crc\" " Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.452678 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbf6t\" (UniqueName: \"kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-kube-api-access-cbf6t\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.452689 4777 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7163366e-9a39-4379-af0a-d6c7c998ab7f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.452697 4777 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.452706 4777 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.452715 4777 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7163366e-9a39-4379-af0a-d6c7c998ab7f-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.452724 4777 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.452731 4777 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.452739 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7163366e-9a39-4379-af0a-d6c7c998ab7f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.481193 4777 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.482011 4777 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f") on node "crc" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.503789 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "7163366e-9a39-4379-af0a-d6c7c998ab7f" (UID: "7163366e-9a39-4379-af0a-d6c7c998ab7f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.554381 4777 reconciler_common.go:293] "Volume detached for volume \"pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:04 crc kubenswrapper[4777]: I1124 17:25:04.554415 4777 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7163366e-9a39-4379-af0a-d6c7c998ab7f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.069100 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7163366e-9a39-4379-af0a-d6c7c998ab7f","Type":"ContainerDied","Data":"eafb2532315b1b71211c72b4081052bcf8ef5a55b922201b10bcf3e1492ba376"} Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.070268 4777 scope.go:117] "RemoveContainer" containerID="de911e01738284bd59a6b0afed7357ce4a53fef2f11ed85c0d07b82a41044436" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.069153 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.075749 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1c366c57-2d20-4e76-b90f-c90fc01347fc","Type":"ContainerStarted","Data":"7eb2fe0c089524d468b8450601f063a498c9cb1ff2151728b2ba19d895781f2c"} Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.075790 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"1c366c57-2d20-4e76-b90f-c90fc01347fc","Type":"ContainerStarted","Data":"6ab73ac5d97ff76c13d6714888a009505734c7851af8e0699b255edf86b220b0"} Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.079335 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5","Type":"ContainerStarted","Data":"c024f7134be089989642602d687e3c22b8ffd6ede187f89925a0f096035b0e13"} Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.079376 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5","Type":"ContainerStarted","Data":"30732c5b6fced348a623563e056ff6ef726b1ec10e716280508feb8ac853d63e"} Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.079389 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"5f4b4358-805a-4a55-a5a8-9200c0c5e5b5","Type":"ContainerStarted","Data":"7394d69308527775cc2f0345920d9ab6f2961a361cbacf442ca3d817dbc96159"} Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.103444 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=1.856694901 podStartE2EDuration="2.103428372s" podCreationTimestamp="2025-11-24 17:25:03 +0000 UTC" firstStartedPulling="2025-11-24 17:25:04.226345879 +0000 UTC m=+1462.385540928" lastFinishedPulling="2025-11-24 17:25:04.47307935 +0000 UTC m=+1462.632274399" observedRunningTime="2025-11-24 17:25:05.100949851 +0000 UTC m=+1463.260144920" watchObservedRunningTime="2025-11-24 17:25:05.103428372 +0000 UTC m=+1463.262623431" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.120016 4777 scope.go:117] "RemoveContainer" containerID="51d916dedea9f0bddc55f0a8a97bf14ba1e65ae0a91b751dee93cb87d18481b8" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.148013 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.168198 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.172205 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=2.172187742 podStartE2EDuration="2.172187742s" podCreationTimestamp="2025-11-24 17:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:25:05.150755592 +0000 UTC m=+1463.309950661" watchObservedRunningTime="2025-11-24 17:25:05.172187742 +0000 UTC m=+1463.331382791" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.191264 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:25:05 crc kubenswrapper[4777]: E1124 17:25:05.191915 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7163366e-9a39-4379-af0a-d6c7c998ab7f" containerName="rabbitmq" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.191943 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="7163366e-9a39-4379-af0a-d6c7c998ab7f" containerName="rabbitmq" Nov 24 17:25:05 crc kubenswrapper[4777]: E1124 17:25:05.192000 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7163366e-9a39-4379-af0a-d6c7c998ab7f" containerName="setup-container" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.192014 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="7163366e-9a39-4379-af0a-d6c7c998ab7f" containerName="setup-container" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.192301 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="7163366e-9a39-4379-af0a-d6c7c998ab7f" containerName="rabbitmq" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.194046 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.201253 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4hmjq" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.201418 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.202876 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.203020 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.203132 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.203243 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.203297 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.203393 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.283321 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7163366e-9a39-4379-af0a-d6c7c998ab7f" path="/var/lib/kubelet/pods/7163366e-9a39-4379-af0a-d6c7c998ab7f/volumes" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.373226 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zr9x\" (UniqueName: \"kubernetes.io/projected/42d2ab97-ca44-4234-8405-22f8797e0c0b-kube-api-access-9zr9x\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.373266 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/42d2ab97-ca44-4234-8405-22f8797e0c0b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.373313 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/42d2ab97-ca44-4234-8405-22f8797e0c0b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.373450 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.373507 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/42d2ab97-ca44-4234-8405-22f8797e0c0b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.373598 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/42d2ab97-ca44-4234-8405-22f8797e0c0b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.373701 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/42d2ab97-ca44-4234-8405-22f8797e0c0b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.373739 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/42d2ab97-ca44-4234-8405-22f8797e0c0b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.373762 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/42d2ab97-ca44-4234-8405-22f8797e0c0b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.373845 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/42d2ab97-ca44-4234-8405-22f8797e0c0b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.374042 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/42d2ab97-ca44-4234-8405-22f8797e0c0b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.475744 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/42d2ab97-ca44-4234-8405-22f8797e0c0b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.475831 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/42d2ab97-ca44-4234-8405-22f8797e0c0b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.475860 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zr9x\" (UniqueName: \"kubernetes.io/projected/42d2ab97-ca44-4234-8405-22f8797e0c0b-kube-api-access-9zr9x\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.475912 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/42d2ab97-ca44-4234-8405-22f8797e0c0b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.475955 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.476009 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/42d2ab97-ca44-4234-8405-22f8797e0c0b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.476063 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/42d2ab97-ca44-4234-8405-22f8797e0c0b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.476103 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/42d2ab97-ca44-4234-8405-22f8797e0c0b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.476128 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/42d2ab97-ca44-4234-8405-22f8797e0c0b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.476151 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/42d2ab97-ca44-4234-8405-22f8797e0c0b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.476190 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/42d2ab97-ca44-4234-8405-22f8797e0c0b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.476766 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/42d2ab97-ca44-4234-8405-22f8797e0c0b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.477377 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/42d2ab97-ca44-4234-8405-22f8797e0c0b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.478618 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/42d2ab97-ca44-4234-8405-22f8797e0c0b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.478623 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/42d2ab97-ca44-4234-8405-22f8797e0c0b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.478927 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/42d2ab97-ca44-4234-8405-22f8797e0c0b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.481790 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/42d2ab97-ca44-4234-8405-22f8797e0c0b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.481838 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/42d2ab97-ca44-4234-8405-22f8797e0c0b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.482267 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/42d2ab97-ca44-4234-8405-22f8797e0c0b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.484954 4777 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.485006 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/64e6484d692ffc427f079db163a6addcadc33dfbfa9e1c0ba7f6f979b99f192e/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.500227 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/42d2ab97-ca44-4234-8405-22f8797e0c0b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.506566 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zr9x\" (UniqueName: \"kubernetes.io/projected/42d2ab97-ca44-4234-8405-22f8797e0c0b-kube-api-access-9zr9x\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.570605 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9028b9c-dbae-4698-ae4f-693adc35d56f\") pod \"rabbitmq-cell1-server-0\" (UID: \"42d2ab97-ca44-4234-8405-22f8797e0c0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:05 crc kubenswrapper[4777]: I1124 17:25:05.873469 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.129807 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.426340 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dc7c944bf-dp4s5"] Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.440719 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.446508 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.469252 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dc7c944bf-dp4s5"] Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.622637 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dc7c944bf-dp4s5"] Nov 24 17:25:06 crc kubenswrapper[4777]: E1124 17:25:06.623911 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-wvdk6 openstack-edpm-ipam ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" podUID="6ff112be-edb4-4ca4-aedf-c178234b7913" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.624363 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-config\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.624434 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-dns-swift-storage-0\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.624462 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-ovsdbserver-sb\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.624489 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-dns-svc\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.624515 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-openstack-edpm-ipam\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.624543 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvdk6\" (UniqueName: \"kubernetes.io/projected/6ff112be-edb4-4ca4-aedf-c178234b7913-kube-api-access-wvdk6\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.624652 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-ovsdbserver-nb\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.642265 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c4b758ff5-frg8p"] Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.644095 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.656732 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c4b758ff5-frg8p"] Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.727768 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-config\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.727839 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-dns-swift-storage-0\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.727867 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-ovsdbserver-sb\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.727900 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-dns-svc\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.727938 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-openstack-edpm-ipam\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.727990 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvdk6\" (UniqueName: \"kubernetes.io/projected/6ff112be-edb4-4ca4-aedf-c178234b7913-kube-api-access-wvdk6\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.728108 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-ovsdbserver-nb\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.729268 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-ovsdbserver-nb\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.729783 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-config\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.730358 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-dns-swift-storage-0\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.730721 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-dns-svc\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.733747 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-ovsdbserver-sb\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.749043 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.760400 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-openstack-edpm-ipam\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.781089 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvdk6\" (UniqueName: \"kubernetes.io/projected/6ff112be-edb4-4ca4-aedf-c178234b7913-kube-api-access-wvdk6\") pod \"dnsmasq-dns-dc7c944bf-dp4s5\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.834168 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-dns-swift-storage-0\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.834555 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-ovsdbserver-nb\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.834577 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-config\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.834618 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-openstack-edpm-ipam\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.834685 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rzqn\" (UniqueName: \"kubernetes.io/projected/33b50ce8-58e3-457c-81ce-dde13962d338-kube-api-access-4rzqn\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.834707 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-ovsdbserver-sb\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.834740 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-dns-svc\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.937163 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-ovsdbserver-nb\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.937212 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-config\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.937260 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-openstack-edpm-ipam\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.937320 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rzqn\" (UniqueName: \"kubernetes.io/projected/33b50ce8-58e3-457c-81ce-dde13962d338-kube-api-access-4rzqn\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.937342 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-ovsdbserver-sb\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.937375 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-dns-svc\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.937430 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-dns-swift-storage-0\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.938167 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-config\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.938416 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-ovsdbserver-nb\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.938583 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-dns-swift-storage-0\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.938714 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-openstack-edpm-ipam\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.938993 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-ovsdbserver-sb\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.939394 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33b50ce8-58e3-457c-81ce-dde13962d338-dns-svc\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.960836 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rzqn\" (UniqueName: \"kubernetes.io/projected/33b50ce8-58e3-457c-81ce-dde13962d338-kube-api-access-4rzqn\") pod \"dnsmasq-dns-c4b758ff5-frg8p\" (UID: \"33b50ce8-58e3-457c-81ce-dde13962d338\") " pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:06 crc kubenswrapper[4777]: I1124 17:25:06.984423 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.149098 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.149698 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"42d2ab97-ca44-4234-8405-22f8797e0c0b","Type":"ContainerStarted","Data":"09e27c7b52b91f4f56a39a9ff9f328f4eecc55e67882f90ae5db4d7a3b2917e4"} Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.171260 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.347677 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvdk6\" (UniqueName: \"kubernetes.io/projected/6ff112be-edb4-4ca4-aedf-c178234b7913-kube-api-access-wvdk6\") pod \"6ff112be-edb4-4ca4-aedf-c178234b7913\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.348052 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-openstack-edpm-ipam\") pod \"6ff112be-edb4-4ca4-aedf-c178234b7913\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.348095 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-ovsdbserver-nb\") pod \"6ff112be-edb4-4ca4-aedf-c178234b7913\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.348189 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-dns-svc\") pod \"6ff112be-edb4-4ca4-aedf-c178234b7913\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.348230 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-ovsdbserver-sb\") pod \"6ff112be-edb4-4ca4-aedf-c178234b7913\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.348344 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-dns-swift-storage-0\") pod \"6ff112be-edb4-4ca4-aedf-c178234b7913\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.348391 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-config\") pod \"6ff112be-edb4-4ca4-aedf-c178234b7913\" (UID: \"6ff112be-edb4-4ca4-aedf-c178234b7913\") " Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.349303 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-config" (OuterVolumeSpecName: "config") pod "6ff112be-edb4-4ca4-aedf-c178234b7913" (UID: "6ff112be-edb4-4ca4-aedf-c178234b7913"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.349358 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "6ff112be-edb4-4ca4-aedf-c178234b7913" (UID: "6ff112be-edb4-4ca4-aedf-c178234b7913"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.349612 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6ff112be-edb4-4ca4-aedf-c178234b7913" (UID: "6ff112be-edb4-4ca4-aedf-c178234b7913"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.349950 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6ff112be-edb4-4ca4-aedf-c178234b7913" (UID: "6ff112be-edb4-4ca4-aedf-c178234b7913"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.350605 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6ff112be-edb4-4ca4-aedf-c178234b7913" (UID: "6ff112be-edb4-4ca4-aedf-c178234b7913"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.351293 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6ff112be-edb4-4ca4-aedf-c178234b7913" (UID: "6ff112be-edb4-4ca4-aedf-c178234b7913"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.374136 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ff112be-edb4-4ca4-aedf-c178234b7913-kube-api-access-wvdk6" (OuterVolumeSpecName: "kube-api-access-wvdk6") pod "6ff112be-edb4-4ca4-aedf-c178234b7913" (UID: "6ff112be-edb4-4ca4-aedf-c178234b7913"). InnerVolumeSpecName "kube-api-access-wvdk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.450660 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvdk6\" (UniqueName: \"kubernetes.io/projected/6ff112be-edb4-4ca4-aedf-c178234b7913-kube-api-access-wvdk6\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.450700 4777 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.450709 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.450718 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.450726 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.450736 4777 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.450744 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ff112be-edb4-4ca4-aedf-c178234b7913-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:07 crc kubenswrapper[4777]: I1124 17:25:07.618492 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c4b758ff5-frg8p"] Nov 24 17:25:08 crc kubenswrapper[4777]: I1124 17:25:08.162654 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc7c944bf-dp4s5" Nov 24 17:25:08 crc kubenswrapper[4777]: I1124 17:25:08.163121 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" event={"ID":"33b50ce8-58e3-457c-81ce-dde13962d338","Type":"ContainerStarted","Data":"5d5a36d3c8f1bf5efdbea6b6dad4d3307350e77fb9def3df3f2147495bdaba26"} Nov 24 17:25:08 crc kubenswrapper[4777]: I1124 17:25:08.236053 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:25:08 crc kubenswrapper[4777]: I1124 17:25:08.255012 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dc7c944bf-dp4s5"] Nov 24 17:25:08 crc kubenswrapper[4777]: I1124 17:25:08.274517 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dc7c944bf-dp4s5"] Nov 24 17:25:08 crc kubenswrapper[4777]: I1124 17:25:08.319252 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:25:08 crc kubenswrapper[4777]: I1124 17:25:08.490749 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h2fbd"] Nov 24 17:25:09 crc kubenswrapper[4777]: I1124 17:25:09.172205 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"42d2ab97-ca44-4234-8405-22f8797e0c0b","Type":"ContainerStarted","Data":"03071f6d86b304c649bc47515f4441f88c94fb5169ab71edb678543725475096"} Nov 24 17:25:09 crc kubenswrapper[4777]: I1124 17:25:09.174110 4777 generic.go:334] "Generic (PLEG): container finished" podID="33b50ce8-58e3-457c-81ce-dde13962d338" containerID="1f7c89828cb623c53c23444bb55b57b28ab2469300b7c013d0d8df72e81abd45" exitCode=0 Nov 24 17:25:09 crc kubenswrapper[4777]: I1124 17:25:09.174199 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" event={"ID":"33b50ce8-58e3-457c-81ce-dde13962d338","Type":"ContainerDied","Data":"1f7c89828cb623c53c23444bb55b57b28ab2469300b7c013d0d8df72e81abd45"} Nov 24 17:25:09 crc kubenswrapper[4777]: I1124 17:25:09.264507 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ff112be-edb4-4ca4-aedf-c178234b7913" path="/var/lib/kubelet/pods/6ff112be-edb4-4ca4-aedf-c178234b7913/volumes" Nov 24 17:25:10 crc kubenswrapper[4777]: I1124 17:25:10.186379 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" event={"ID":"33b50ce8-58e3-457c-81ce-dde13962d338","Type":"ContainerStarted","Data":"b6785b8fa4c57ded4b78bda617eee5ffed32c481ac5a23746481d2552c80f8e5"} Nov 24 17:25:10 crc kubenswrapper[4777]: I1124 17:25:10.186510 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h2fbd" podUID="e36159e1-7be5-49a7-8417-57253b4e7be0" containerName="registry-server" containerID="cri-o://5d33d49edaff9e300c7160061f0be9fba9b30177e288a940b0e597f28f13ad06" gracePeriod=2 Nov 24 17:25:10 crc kubenswrapper[4777]: I1124 17:25:10.219674 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" podStartSLOduration=4.219472432 podStartE2EDuration="4.219472432s" podCreationTimestamp="2025-11-24 17:25:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:25:10.209064281 +0000 UTC m=+1468.368259320" watchObservedRunningTime="2025-11-24 17:25:10.219472432 +0000 UTC m=+1468.378667481" Nov 24 17:25:10 crc kubenswrapper[4777]: I1124 17:25:10.814595 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:25:10 crc kubenswrapper[4777]: I1124 17:25:10.923420 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e36159e1-7be5-49a7-8417-57253b4e7be0-utilities\") pod \"e36159e1-7be5-49a7-8417-57253b4e7be0\" (UID: \"e36159e1-7be5-49a7-8417-57253b4e7be0\") " Nov 24 17:25:10 crc kubenswrapper[4777]: I1124 17:25:10.923502 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e36159e1-7be5-49a7-8417-57253b4e7be0-catalog-content\") pod \"e36159e1-7be5-49a7-8417-57253b4e7be0\" (UID: \"e36159e1-7be5-49a7-8417-57253b4e7be0\") " Nov 24 17:25:10 crc kubenswrapper[4777]: I1124 17:25:10.923619 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kvj5\" (UniqueName: \"kubernetes.io/projected/e36159e1-7be5-49a7-8417-57253b4e7be0-kube-api-access-7kvj5\") pod \"e36159e1-7be5-49a7-8417-57253b4e7be0\" (UID: \"e36159e1-7be5-49a7-8417-57253b4e7be0\") " Nov 24 17:25:10 crc kubenswrapper[4777]: I1124 17:25:10.924494 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e36159e1-7be5-49a7-8417-57253b4e7be0-utilities" (OuterVolumeSpecName: "utilities") pod "e36159e1-7be5-49a7-8417-57253b4e7be0" (UID: "e36159e1-7be5-49a7-8417-57253b4e7be0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:25:10 crc kubenswrapper[4777]: I1124 17:25:10.933392 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e36159e1-7be5-49a7-8417-57253b4e7be0-kube-api-access-7kvj5" (OuterVolumeSpecName: "kube-api-access-7kvj5") pod "e36159e1-7be5-49a7-8417-57253b4e7be0" (UID: "e36159e1-7be5-49a7-8417-57253b4e7be0"). InnerVolumeSpecName "kube-api-access-7kvj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.026955 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kvj5\" (UniqueName: \"kubernetes.io/projected/e36159e1-7be5-49a7-8417-57253b4e7be0-kube-api-access-7kvj5\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.027012 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e36159e1-7be5-49a7-8417-57253b4e7be0-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.028357 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e36159e1-7be5-49a7-8417-57253b4e7be0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e36159e1-7be5-49a7-8417-57253b4e7be0" (UID: "e36159e1-7be5-49a7-8417-57253b4e7be0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.130037 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e36159e1-7be5-49a7-8417-57253b4e7be0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.199844 4777 generic.go:334] "Generic (PLEG): container finished" podID="e36159e1-7be5-49a7-8417-57253b4e7be0" containerID="5d33d49edaff9e300c7160061f0be9fba9b30177e288a940b0e597f28f13ad06" exitCode=0 Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.199888 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h2fbd" event={"ID":"e36159e1-7be5-49a7-8417-57253b4e7be0","Type":"ContainerDied","Data":"5d33d49edaff9e300c7160061f0be9fba9b30177e288a940b0e597f28f13ad06"} Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.199946 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h2fbd" Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.199981 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h2fbd" event={"ID":"e36159e1-7be5-49a7-8417-57253b4e7be0","Type":"ContainerDied","Data":"a0f2954607e8bcf3f1a5629142d5a65b7e1d0786518f39b33fef13d59889f470"} Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.200009 4777 scope.go:117] "RemoveContainer" containerID="5d33d49edaff9e300c7160061f0be9fba9b30177e288a940b0e597f28f13ad06" Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.200151 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.229389 4777 scope.go:117] "RemoveContainer" containerID="98605f3510bf5886f3d7ec67fb9f80aea188b7a6172441ca71246375c34ba891" Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.268828 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h2fbd"] Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.268879 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h2fbd"] Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.268982 4777 scope.go:117] "RemoveContainer" containerID="ab7e9e98a34b1988684706f4ddeaf816d942278c54ee4a8d6a45503c68ca6d2c" Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.309344 4777 scope.go:117] "RemoveContainer" containerID="5d33d49edaff9e300c7160061f0be9fba9b30177e288a940b0e597f28f13ad06" Nov 24 17:25:11 crc kubenswrapper[4777]: E1124 17:25:11.309903 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d33d49edaff9e300c7160061f0be9fba9b30177e288a940b0e597f28f13ad06\": container with ID starting with 5d33d49edaff9e300c7160061f0be9fba9b30177e288a940b0e597f28f13ad06 not found: ID does not exist" containerID="5d33d49edaff9e300c7160061f0be9fba9b30177e288a940b0e597f28f13ad06" Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.309954 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d33d49edaff9e300c7160061f0be9fba9b30177e288a940b0e597f28f13ad06"} err="failed to get container status \"5d33d49edaff9e300c7160061f0be9fba9b30177e288a940b0e597f28f13ad06\": rpc error: code = NotFound desc = could not find container \"5d33d49edaff9e300c7160061f0be9fba9b30177e288a940b0e597f28f13ad06\": container with ID starting with 5d33d49edaff9e300c7160061f0be9fba9b30177e288a940b0e597f28f13ad06 not found: ID does not exist" Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.310001 4777 scope.go:117] "RemoveContainer" containerID="98605f3510bf5886f3d7ec67fb9f80aea188b7a6172441ca71246375c34ba891" Nov 24 17:25:11 crc kubenswrapper[4777]: E1124 17:25:11.310497 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98605f3510bf5886f3d7ec67fb9f80aea188b7a6172441ca71246375c34ba891\": container with ID starting with 98605f3510bf5886f3d7ec67fb9f80aea188b7a6172441ca71246375c34ba891 not found: ID does not exist" containerID="98605f3510bf5886f3d7ec67fb9f80aea188b7a6172441ca71246375c34ba891" Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.310530 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98605f3510bf5886f3d7ec67fb9f80aea188b7a6172441ca71246375c34ba891"} err="failed to get container status \"98605f3510bf5886f3d7ec67fb9f80aea188b7a6172441ca71246375c34ba891\": rpc error: code = NotFound desc = could not find container \"98605f3510bf5886f3d7ec67fb9f80aea188b7a6172441ca71246375c34ba891\": container with ID starting with 98605f3510bf5886f3d7ec67fb9f80aea188b7a6172441ca71246375c34ba891 not found: ID does not exist" Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.310559 4777 scope.go:117] "RemoveContainer" containerID="ab7e9e98a34b1988684706f4ddeaf816d942278c54ee4a8d6a45503c68ca6d2c" Nov 24 17:25:11 crc kubenswrapper[4777]: E1124 17:25:11.310872 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab7e9e98a34b1988684706f4ddeaf816d942278c54ee4a8d6a45503c68ca6d2c\": container with ID starting with ab7e9e98a34b1988684706f4ddeaf816d942278c54ee4a8d6a45503c68ca6d2c not found: ID does not exist" containerID="ab7e9e98a34b1988684706f4ddeaf816d942278c54ee4a8d6a45503c68ca6d2c" Nov 24 17:25:11 crc kubenswrapper[4777]: I1124 17:25:11.310905 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab7e9e98a34b1988684706f4ddeaf816d942278c54ee4a8d6a45503c68ca6d2c"} err="failed to get container status \"ab7e9e98a34b1988684706f4ddeaf816d942278c54ee4a8d6a45503c68ca6d2c\": rpc error: code = NotFound desc = could not find container \"ab7e9e98a34b1988684706f4ddeaf816d942278c54ee4a8d6a45503c68ca6d2c\": container with ID starting with ab7e9e98a34b1988684706f4ddeaf816d942278c54ee4a8d6a45503c68ca6d2c not found: ID does not exist" Nov 24 17:25:13 crc kubenswrapper[4777]: I1124 17:25:13.268865 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e36159e1-7be5-49a7-8417-57253b4e7be0" path="/var/lib/kubelet/pods/e36159e1-7be5-49a7-8417-57253b4e7be0/volumes" Nov 24 17:25:14 crc kubenswrapper[4777]: I1124 17:25:14.270041 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:25:14 crc kubenswrapper[4777]: I1124 17:25:14.270098 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:25:16 crc kubenswrapper[4777]: I1124 17:25:16.986862 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-c4b758ff5-frg8p" Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.068176 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54dd998c-8smx5"] Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.068453 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54dd998c-8smx5" podUID="4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4" containerName="dnsmasq-dns" containerID="cri-o://329084f1a1bf025edf20124bb61ab4512356cda5cfe18ffc22697058d48ceff6" gracePeriod=10 Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.288832 4777 generic.go:334] "Generic (PLEG): container finished" podID="4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4" containerID="329084f1a1bf025edf20124bb61ab4512356cda5cfe18ffc22697058d48ceff6" exitCode=0 Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.288867 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54dd998c-8smx5" event={"ID":"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4","Type":"ContainerDied","Data":"329084f1a1bf025edf20124bb61ab4512356cda5cfe18ffc22697058d48ceff6"} Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.632349 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.705460 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-ovsdbserver-nb\") pod \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.705689 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-config\") pod \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.705749 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddht8\" (UniqueName: \"kubernetes.io/projected/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-kube-api-access-ddht8\") pod \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.705795 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-dns-svc\") pod \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.705894 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-dns-swift-storage-0\") pod \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.706054 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-ovsdbserver-sb\") pod \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\" (UID: \"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4\") " Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.712566 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-kube-api-access-ddht8" (OuterVolumeSpecName: "kube-api-access-ddht8") pod "4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4" (UID: "4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4"). InnerVolumeSpecName "kube-api-access-ddht8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.774987 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4" (UID: "4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.781431 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4" (UID: "4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.783263 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-config" (OuterVolumeSpecName: "config") pod "4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4" (UID: "4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.785691 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4" (UID: "4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.790804 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4" (UID: "4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.808784 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.808816 4777 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.808829 4777 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.808841 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddht8\" (UniqueName: \"kubernetes.io/projected/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-kube-api-access-ddht8\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.808853 4777 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:17 crc kubenswrapper[4777]: I1124 17:25:17.808863 4777 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:18 crc kubenswrapper[4777]: I1124 17:25:18.303078 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54dd998c-8smx5" event={"ID":"4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4","Type":"ContainerDied","Data":"eb78c2d9a75d4d54545968a46cc226dcf857934c931503744f2920b95884b23e"} Nov 24 17:25:18 crc kubenswrapper[4777]: I1124 17:25:18.303128 4777 scope.go:117] "RemoveContainer" containerID="329084f1a1bf025edf20124bb61ab4512356cda5cfe18ffc22697058d48ceff6" Nov 24 17:25:18 crc kubenswrapper[4777]: I1124 17:25:18.303202 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54dd998c-8smx5" Nov 24 17:25:18 crc kubenswrapper[4777]: I1124 17:25:18.329375 4777 scope.go:117] "RemoveContainer" containerID="8c327a473dbd138850bbd3af0d40527f3cba217e1499d8a8bc72b9e6602db3ec" Nov 24 17:25:18 crc kubenswrapper[4777]: I1124 17:25:18.348793 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54dd998c-8smx5"] Nov 24 17:25:18 crc kubenswrapper[4777]: I1124 17:25:18.357771 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54dd998c-8smx5"] Nov 24 17:25:19 crc kubenswrapper[4777]: I1124 17:25:19.265793 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4" path="/var/lib/kubelet/pods/4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4/volumes" Nov 24 17:25:25 crc kubenswrapper[4777]: I1124 17:25:25.172327 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 17:25:25 crc kubenswrapper[4777]: I1124 17:25:25.991247 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq"] Nov 24 17:25:25 crc kubenswrapper[4777]: E1124 17:25:25.992196 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4" containerName="dnsmasq-dns" Nov 24 17:25:25 crc kubenswrapper[4777]: I1124 17:25:25.992221 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4" containerName="dnsmasq-dns" Nov 24 17:25:25 crc kubenswrapper[4777]: E1124 17:25:25.992245 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e36159e1-7be5-49a7-8417-57253b4e7be0" containerName="extract-content" Nov 24 17:25:25 crc kubenswrapper[4777]: I1124 17:25:25.992254 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="e36159e1-7be5-49a7-8417-57253b4e7be0" containerName="extract-content" Nov 24 17:25:25 crc kubenswrapper[4777]: E1124 17:25:25.992271 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4" containerName="init" Nov 24 17:25:25 crc kubenswrapper[4777]: I1124 17:25:25.992278 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4" containerName="init" Nov 24 17:25:25 crc kubenswrapper[4777]: E1124 17:25:25.992290 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e36159e1-7be5-49a7-8417-57253b4e7be0" containerName="registry-server" Nov 24 17:25:25 crc kubenswrapper[4777]: I1124 17:25:25.992298 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="e36159e1-7be5-49a7-8417-57253b4e7be0" containerName="registry-server" Nov 24 17:25:25 crc kubenswrapper[4777]: E1124 17:25:25.992338 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e36159e1-7be5-49a7-8417-57253b4e7be0" containerName="extract-utilities" Nov 24 17:25:25 crc kubenswrapper[4777]: I1124 17:25:25.992345 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="e36159e1-7be5-49a7-8417-57253b4e7be0" containerName="extract-utilities" Nov 24 17:25:25 crc kubenswrapper[4777]: I1124 17:25:25.992602 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d5753b3-7c49-4c92-9f1b-cf5ada5f28b4" containerName="dnsmasq-dns" Nov 24 17:25:25 crc kubenswrapper[4777]: I1124 17:25:25.992625 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="e36159e1-7be5-49a7-8417-57253b4e7be0" containerName="registry-server" Nov 24 17:25:25 crc kubenswrapper[4777]: I1124 17:25:25.993448 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:25 crc kubenswrapper[4777]: I1124 17:25:25.995649 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:25:25 crc kubenswrapper[4777]: I1124 17:25:25.996498 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:25:25 crc kubenswrapper[4777]: I1124 17:25:25.996715 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:25:26 crc kubenswrapper[4777]: I1124 17:25:26.001089 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:25:26 crc kubenswrapper[4777]: I1124 17:25:26.001453 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq"] Nov 24 17:25:26 crc kubenswrapper[4777]: I1124 17:25:26.101465 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7pjs\" (UniqueName: \"kubernetes.io/projected/10de4ea3-2dc3-4e57-ac55-d6be12743347-kube-api-access-c7pjs\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:26 crc kubenswrapper[4777]: I1124 17:25:26.101700 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:26 crc kubenswrapper[4777]: I1124 17:25:26.101812 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:26 crc kubenswrapper[4777]: I1124 17:25:26.101836 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:26 crc kubenswrapper[4777]: I1124 17:25:26.203431 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:26 crc kubenswrapper[4777]: I1124 17:25:26.203484 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:26 crc kubenswrapper[4777]: I1124 17:25:26.203660 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7pjs\" (UniqueName: \"kubernetes.io/projected/10de4ea3-2dc3-4e57-ac55-d6be12743347-kube-api-access-c7pjs\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:26 crc kubenswrapper[4777]: I1124 17:25:26.203730 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:26 crc kubenswrapper[4777]: I1124 17:25:26.208221 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:26 crc kubenswrapper[4777]: I1124 17:25:26.208690 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:26 crc kubenswrapper[4777]: I1124 17:25:26.211388 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:26 crc kubenswrapper[4777]: I1124 17:25:26.220990 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7pjs\" (UniqueName: \"kubernetes.io/projected/10de4ea3-2dc3-4e57-ac55-d6be12743347-kube-api-access-c7pjs\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:26 crc kubenswrapper[4777]: I1124 17:25:26.312830 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:27 crc kubenswrapper[4777]: I1124 17:25:27.344234 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq"] Nov 24 17:25:27 crc kubenswrapper[4777]: I1124 17:25:27.430444 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" event={"ID":"10de4ea3-2dc3-4e57-ac55-d6be12743347","Type":"ContainerStarted","Data":"c6247d9f944f48910da2fb03eb373c525022dcd3ca7de93fdb25332f6c9798a1"} Nov 24 17:25:31 crc kubenswrapper[4777]: I1124 17:25:31.156505 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-28c29"] Nov 24 17:25:31 crc kubenswrapper[4777]: I1124 17:25:31.159066 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:31 crc kubenswrapper[4777]: I1124 17:25:31.169083 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-28c29"] Nov 24 17:25:31 crc kubenswrapper[4777]: I1124 17:25:31.210138 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-catalog-content\") pod \"redhat-marketplace-28c29\" (UID: \"29d9b08e-8a4b-4e0b-9b83-724b303e2e61\") " pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:31 crc kubenswrapper[4777]: I1124 17:25:31.210502 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-utilities\") pod \"redhat-marketplace-28c29\" (UID: \"29d9b08e-8a4b-4e0b-9b83-724b303e2e61\") " pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:31 crc kubenswrapper[4777]: I1124 17:25:31.210663 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5knll\" (UniqueName: \"kubernetes.io/projected/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-kube-api-access-5knll\") pod \"redhat-marketplace-28c29\" (UID: \"29d9b08e-8a4b-4e0b-9b83-724b303e2e61\") " pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:31 crc kubenswrapper[4777]: I1124 17:25:31.312196 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-utilities\") pod \"redhat-marketplace-28c29\" (UID: \"29d9b08e-8a4b-4e0b-9b83-724b303e2e61\") " pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:31 crc kubenswrapper[4777]: I1124 17:25:31.312351 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5knll\" (UniqueName: \"kubernetes.io/projected/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-kube-api-access-5knll\") pod \"redhat-marketplace-28c29\" (UID: \"29d9b08e-8a4b-4e0b-9b83-724b303e2e61\") " pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:31 crc kubenswrapper[4777]: I1124 17:25:31.312405 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-catalog-content\") pod \"redhat-marketplace-28c29\" (UID: \"29d9b08e-8a4b-4e0b-9b83-724b303e2e61\") " pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:31 crc kubenswrapper[4777]: I1124 17:25:31.312915 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-utilities\") pod \"redhat-marketplace-28c29\" (UID: \"29d9b08e-8a4b-4e0b-9b83-724b303e2e61\") " pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:31 crc kubenswrapper[4777]: I1124 17:25:31.312952 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-catalog-content\") pod \"redhat-marketplace-28c29\" (UID: \"29d9b08e-8a4b-4e0b-9b83-724b303e2e61\") " pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:31 crc kubenswrapper[4777]: I1124 17:25:31.330452 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5knll\" (UniqueName: \"kubernetes.io/projected/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-kube-api-access-5knll\") pod \"redhat-marketplace-28c29\" (UID: \"29d9b08e-8a4b-4e0b-9b83-724b303e2e61\") " pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:31 crc kubenswrapper[4777]: I1124 17:25:31.487404 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:33 crc kubenswrapper[4777]: I1124 17:25:33.548238 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-np6rr"] Nov 24 17:25:33 crc kubenswrapper[4777]: I1124 17:25:33.551759 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:25:33 crc kubenswrapper[4777]: I1124 17:25:33.577036 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-np6rr"] Nov 24 17:25:33 crc kubenswrapper[4777]: I1124 17:25:33.678841 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-utilities\") pod \"community-operators-np6rr\" (UID: \"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a\") " pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:25:33 crc kubenswrapper[4777]: I1124 17:25:33.679208 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2w7g\" (UniqueName: \"kubernetes.io/projected/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-kube-api-access-g2w7g\") pod \"community-operators-np6rr\" (UID: \"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a\") " pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:25:33 crc kubenswrapper[4777]: I1124 17:25:33.679648 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-catalog-content\") pod \"community-operators-np6rr\" (UID: \"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a\") " pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:25:33 crc kubenswrapper[4777]: I1124 17:25:33.781331 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-catalog-content\") pod \"community-operators-np6rr\" (UID: \"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a\") " pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:25:33 crc kubenswrapper[4777]: I1124 17:25:33.782101 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-utilities\") pod \"community-operators-np6rr\" (UID: \"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a\") " pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:25:33 crc kubenswrapper[4777]: I1124 17:25:33.782296 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2w7g\" (UniqueName: \"kubernetes.io/projected/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-kube-api-access-g2w7g\") pod \"community-operators-np6rr\" (UID: \"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a\") " pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:25:33 crc kubenswrapper[4777]: I1124 17:25:33.782581 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-catalog-content\") pod \"community-operators-np6rr\" (UID: \"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a\") " pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:25:33 crc kubenswrapper[4777]: I1124 17:25:33.782639 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-utilities\") pod \"community-operators-np6rr\" (UID: \"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a\") " pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:25:33 crc kubenswrapper[4777]: I1124 17:25:33.801544 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2w7g\" (UniqueName: \"kubernetes.io/projected/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-kube-api-access-g2w7g\") pod \"community-operators-np6rr\" (UID: \"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a\") " pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:25:33 crc kubenswrapper[4777]: I1124 17:25:33.888388 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:25:36 crc kubenswrapper[4777]: I1124 17:25:36.549007 4777 generic.go:334] "Generic (PLEG): container finished" podID="78712b89-27c4-4481-b152-320fb7fee9fc" containerID="bf8d665b16888241cb30e97b5e1fe0a688ac6405fe01012e413ef52b2e3c76cb" exitCode=0 Nov 24 17:25:36 crc kubenswrapper[4777]: I1124 17:25:36.549104 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"78712b89-27c4-4481-b152-320fb7fee9fc","Type":"ContainerDied","Data":"bf8d665b16888241cb30e97b5e1fe0a688ac6405fe01012e413ef52b2e3c76cb"} Nov 24 17:25:37 crc kubenswrapper[4777]: I1124 17:25:37.406036 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-np6rr"] Nov 24 17:25:37 crc kubenswrapper[4777]: W1124 17:25:37.412868 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5ec65c8_b4cd_4917_bdc0_533df6b58c0a.slice/crio-be4ed43cf96273ccd81fa07484b867aeab3248b7d856fc0669198f45e9b25592 WatchSource:0}: Error finding container be4ed43cf96273ccd81fa07484b867aeab3248b7d856fc0669198f45e9b25592: Status 404 returned error can't find the container with id be4ed43cf96273ccd81fa07484b867aeab3248b7d856fc0669198f45e9b25592 Nov 24 17:25:37 crc kubenswrapper[4777]: I1124 17:25:37.414805 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-28c29"] Nov 24 17:25:37 crc kubenswrapper[4777]: I1124 17:25:37.562312 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-np6rr" event={"ID":"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a","Type":"ContainerStarted","Data":"be4ed43cf96273ccd81fa07484b867aeab3248b7d856fc0669198f45e9b25592"} Nov 24 17:25:37 crc kubenswrapper[4777]: I1124 17:25:37.564282 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"78712b89-27c4-4481-b152-320fb7fee9fc","Type":"ContainerStarted","Data":"9f8f27cf5056a52f0985ad806eb14d5ba06ddf0c729c66bc44719300bae96b05"} Nov 24 17:25:37 crc kubenswrapper[4777]: I1124 17:25:37.564464 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 17:25:37 crc kubenswrapper[4777]: I1124 17:25:37.565865 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" event={"ID":"10de4ea3-2dc3-4e57-ac55-d6be12743347","Type":"ContainerStarted","Data":"aeef61dffee83761be82d9595ed37dd43460bcd67654a8f0b6e1a73cb224cc36"} Nov 24 17:25:37 crc kubenswrapper[4777]: I1124 17:25:37.567355 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28c29" event={"ID":"29d9b08e-8a4b-4e0b-9b83-724b303e2e61","Type":"ContainerStarted","Data":"799dd0e3391e59046adc8ee6250418d61567ee4bb40aa8029da6eaddd7547393"} Nov 24 17:25:37 crc kubenswrapper[4777]: I1124 17:25:37.587487 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.587467116 podStartE2EDuration="36.587467116s" podCreationTimestamp="2025-11-24 17:25:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:25:37.58172128 +0000 UTC m=+1495.740916329" watchObservedRunningTime="2025-11-24 17:25:37.587467116 +0000 UTC m=+1495.746662165" Nov 24 17:25:38 crc kubenswrapper[4777]: I1124 17:25:38.579606 4777 generic.go:334] "Generic (PLEG): container finished" podID="29d9b08e-8a4b-4e0b-9b83-724b303e2e61" containerID="a342a44e213990dd559f9ae6642b7599b942cd3d74840f61260e89ed4acfb4c1" exitCode=0 Nov 24 17:25:38 crc kubenswrapper[4777]: I1124 17:25:38.579663 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28c29" event={"ID":"29d9b08e-8a4b-4e0b-9b83-724b303e2e61","Type":"ContainerDied","Data":"a342a44e213990dd559f9ae6642b7599b942cd3d74840f61260e89ed4acfb4c1"} Nov 24 17:25:38 crc kubenswrapper[4777]: I1124 17:25:38.581888 4777 generic.go:334] "Generic (PLEG): container finished" podID="a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" containerID="fdd99f8b94c8650213b66b985a0fc99847836973cf078e720ba66c0fad66ced4" exitCode=0 Nov 24 17:25:38 crc kubenswrapper[4777]: I1124 17:25:38.581973 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-np6rr" event={"ID":"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a","Type":"ContainerDied","Data":"fdd99f8b94c8650213b66b985a0fc99847836973cf078e720ba66c0fad66ced4"} Nov 24 17:25:38 crc kubenswrapper[4777]: I1124 17:25:38.612199 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" podStartSLOduration=3.946240845 podStartE2EDuration="13.612179511s" podCreationTimestamp="2025-11-24 17:25:25 +0000 UTC" firstStartedPulling="2025-11-24 17:25:27.35278849 +0000 UTC m=+1485.511983549" lastFinishedPulling="2025-11-24 17:25:37.018727166 +0000 UTC m=+1495.177922215" observedRunningTime="2025-11-24 17:25:37.608510095 +0000 UTC m=+1495.767705144" watchObservedRunningTime="2025-11-24 17:25:38.612179511 +0000 UTC m=+1496.771374570" Nov 24 17:25:39 crc kubenswrapper[4777]: I1124 17:25:39.597390 4777 generic.go:334] "Generic (PLEG): container finished" podID="29d9b08e-8a4b-4e0b-9b83-724b303e2e61" containerID="ad0dc0bb5c166a1aaab711bd7499b7809107815948f8a5b52a2c335d91c04768" exitCode=0 Nov 24 17:25:39 crc kubenswrapper[4777]: I1124 17:25:39.597648 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28c29" event={"ID":"29d9b08e-8a4b-4e0b-9b83-724b303e2e61","Type":"ContainerDied","Data":"ad0dc0bb5c166a1aaab711bd7499b7809107815948f8a5b52a2c335d91c04768"} Nov 24 17:25:39 crc kubenswrapper[4777]: I1124 17:25:39.601217 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-np6rr" event={"ID":"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a","Type":"ContainerStarted","Data":"b477dfa57e59c709c249d2391554690f6de794a353d2763eb31600c3a316397b"} Nov 24 17:25:40 crc kubenswrapper[4777]: I1124 17:25:40.612176 4777 generic.go:334] "Generic (PLEG): container finished" podID="a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" containerID="b477dfa57e59c709c249d2391554690f6de794a353d2763eb31600c3a316397b" exitCode=0 Nov 24 17:25:40 crc kubenswrapper[4777]: I1124 17:25:40.612447 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-np6rr" event={"ID":"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a","Type":"ContainerDied","Data":"b477dfa57e59c709c249d2391554690f6de794a353d2763eb31600c3a316397b"} Nov 24 17:25:41 crc kubenswrapper[4777]: I1124 17:25:41.624624 4777 generic.go:334] "Generic (PLEG): container finished" podID="42d2ab97-ca44-4234-8405-22f8797e0c0b" containerID="03071f6d86b304c649bc47515f4441f88c94fb5169ab71edb678543725475096" exitCode=0 Nov 24 17:25:41 crc kubenswrapper[4777]: I1124 17:25:41.624673 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"42d2ab97-ca44-4234-8405-22f8797e0c0b","Type":"ContainerDied","Data":"03071f6d86b304c649bc47515f4441f88c94fb5169ab71edb678543725475096"} Nov 24 17:25:41 crc kubenswrapper[4777]: I1124 17:25:41.744101 4777 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:25:42 crc kubenswrapper[4777]: I1124 17:25:42.008085 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-api-0" Nov 24 17:25:42 crc kubenswrapper[4777]: I1124 17:25:42.635838 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"42d2ab97-ca44-4234-8405-22f8797e0c0b","Type":"ContainerStarted","Data":"b8580d0e5dd007fdfa1fc6021c38666d5eb90a100d732843fba22689fdff2991"} Nov 24 17:25:42 crc kubenswrapper[4777]: I1124 17:25:42.637711 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:42 crc kubenswrapper[4777]: I1124 17:25:42.640130 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28c29" event={"ID":"29d9b08e-8a4b-4e0b-9b83-724b303e2e61","Type":"ContainerStarted","Data":"1ad3184a6047d6ef934e6204aab80fb218594a62fdc8825dd10b23c6908aa3a7"} Nov 24 17:25:42 crc kubenswrapper[4777]: I1124 17:25:42.642260 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-np6rr" event={"ID":"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a","Type":"ContainerStarted","Data":"892bffab92874cd3da4b34c52cc1f57b8f17d8db9e79e13dc1cc58ee82d43d5f"} Nov 24 17:25:42 crc kubenswrapper[4777]: I1124 17:25:42.667207 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.667182395 podStartE2EDuration="37.667182395s" podCreationTimestamp="2025-11-24 17:25:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:25:42.65699863 +0000 UTC m=+1500.816193719" watchObservedRunningTime="2025-11-24 17:25:42.667182395 +0000 UTC m=+1500.826377454" Nov 24 17:25:42 crc kubenswrapper[4777]: I1124 17:25:42.686077 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-28c29" podStartSLOduration=8.411632707 podStartE2EDuration="11.686051291s" podCreationTimestamp="2025-11-24 17:25:31 +0000 UTC" firstStartedPulling="2025-11-24 17:25:38.584575362 +0000 UTC m=+1496.743770421" lastFinishedPulling="2025-11-24 17:25:41.858993956 +0000 UTC m=+1500.018189005" observedRunningTime="2025-11-24 17:25:42.680092729 +0000 UTC m=+1500.839287838" watchObservedRunningTime="2025-11-24 17:25:42.686051291 +0000 UTC m=+1500.845246340" Nov 24 17:25:43 crc kubenswrapper[4777]: I1124 17:25:43.675327 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-np6rr" podStartSLOduration=7.053621287 podStartE2EDuration="10.675307151s" podCreationTimestamp="2025-11-24 17:25:33 +0000 UTC" firstStartedPulling="2025-11-24 17:25:38.583615824 +0000 UTC m=+1496.742810873" lastFinishedPulling="2025-11-24 17:25:42.205301688 +0000 UTC m=+1500.364496737" observedRunningTime="2025-11-24 17:25:43.670139722 +0000 UTC m=+1501.829334781" watchObservedRunningTime="2025-11-24 17:25:43.675307151 +0000 UTC m=+1501.834502210" Nov 24 17:25:43 crc kubenswrapper[4777]: I1124 17:25:43.889259 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:25:43 crc kubenswrapper[4777]: I1124 17:25:43.889562 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:25:44 crc kubenswrapper[4777]: I1124 17:25:44.270641 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:25:44 crc kubenswrapper[4777]: I1124 17:25:44.270697 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:25:44 crc kubenswrapper[4777]: I1124 17:25:44.949155 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-np6rr" podUID="a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" containerName="registry-server" probeResult="failure" output=< Nov 24 17:25:44 crc kubenswrapper[4777]: timeout: failed to connect service ":50051" within 1s Nov 24 17:25:44 crc kubenswrapper[4777]: > Nov 24 17:25:50 crc kubenswrapper[4777]: I1124 17:25:50.761585 4777 generic.go:334] "Generic (PLEG): container finished" podID="10de4ea3-2dc3-4e57-ac55-d6be12743347" containerID="aeef61dffee83761be82d9595ed37dd43460bcd67654a8f0b6e1a73cb224cc36" exitCode=0 Nov 24 17:25:50 crc kubenswrapper[4777]: I1124 17:25:50.762339 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" event={"ID":"10de4ea3-2dc3-4e57-ac55-d6be12743347","Type":"ContainerDied","Data":"aeef61dffee83761be82d9595ed37dd43460bcd67654a8f0b6e1a73cb224cc36"} Nov 24 17:25:51 crc kubenswrapper[4777]: I1124 17:25:51.487710 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:51 crc kubenswrapper[4777]: I1124 17:25:51.488142 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:51 crc kubenswrapper[4777]: I1124 17:25:51.559569 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:51 crc kubenswrapper[4777]: I1124 17:25:51.651188 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 17:25:51 crc kubenswrapper[4777]: I1124 17:25:51.850023 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:51 crc kubenswrapper[4777]: I1124 17:25:51.917474 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-28c29"] Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.309898 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.337509 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-ssh-key\") pod \"10de4ea3-2dc3-4e57-ac55-d6be12743347\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.337862 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-inventory\") pod \"10de4ea3-2dc3-4e57-ac55-d6be12743347\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.338026 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7pjs\" (UniqueName: \"kubernetes.io/projected/10de4ea3-2dc3-4e57-ac55-d6be12743347-kube-api-access-c7pjs\") pod \"10de4ea3-2dc3-4e57-ac55-d6be12743347\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.338160 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-repo-setup-combined-ca-bundle\") pod \"10de4ea3-2dc3-4e57-ac55-d6be12743347\" (UID: \"10de4ea3-2dc3-4e57-ac55-d6be12743347\") " Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.345110 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "10de4ea3-2dc3-4e57-ac55-d6be12743347" (UID: "10de4ea3-2dc3-4e57-ac55-d6be12743347"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.359280 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10de4ea3-2dc3-4e57-ac55-d6be12743347-kube-api-access-c7pjs" (OuterVolumeSpecName: "kube-api-access-c7pjs") pod "10de4ea3-2dc3-4e57-ac55-d6be12743347" (UID: "10de4ea3-2dc3-4e57-ac55-d6be12743347"). InnerVolumeSpecName "kube-api-access-c7pjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.368029 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "10de4ea3-2dc3-4e57-ac55-d6be12743347" (UID: "10de4ea3-2dc3-4e57-ac55-d6be12743347"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.371098 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-inventory" (OuterVolumeSpecName: "inventory") pod "10de4ea3-2dc3-4e57-ac55-d6be12743347" (UID: "10de4ea3-2dc3-4e57-ac55-d6be12743347"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.441059 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.441101 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.441114 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7pjs\" (UniqueName: \"kubernetes.io/projected/10de4ea3-2dc3-4e57-ac55-d6be12743347-kube-api-access-c7pjs\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.441130 4777 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10de4ea3-2dc3-4e57-ac55-d6be12743347-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.788559 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.791304 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq" event={"ID":"10de4ea3-2dc3-4e57-ac55-d6be12743347","Type":"ContainerDied","Data":"c6247d9f944f48910da2fb03eb373c525022dcd3ca7de93fdb25332f6c9798a1"} Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.791372 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6247d9f944f48910da2fb03eb373c525022dcd3ca7de93fdb25332f6c9798a1" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.891917 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5"] Nov 24 17:25:52 crc kubenswrapper[4777]: E1124 17:25:52.892558 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10de4ea3-2dc3-4e57-ac55-d6be12743347" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.892577 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="10de4ea3-2dc3-4e57-ac55-d6be12743347" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.892839 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="10de4ea3-2dc3-4e57-ac55-d6be12743347" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.893793 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.898186 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.899281 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.900002 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.900039 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.904155 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5"] Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.951715 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f27aca44-4a34-469c-97fd-075eed1300dd-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qfmq5\" (UID: \"f27aca44-4a34-469c-97fd-075eed1300dd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.951912 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f27aca44-4a34-469c-97fd-075eed1300dd-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qfmq5\" (UID: \"f27aca44-4a34-469c-97fd-075eed1300dd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" Nov 24 17:25:52 crc kubenswrapper[4777]: I1124 17:25:52.952017 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2rxw\" (UniqueName: \"kubernetes.io/projected/f27aca44-4a34-469c-97fd-075eed1300dd-kube-api-access-m2rxw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qfmq5\" (UID: \"f27aca44-4a34-469c-97fd-075eed1300dd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" Nov 24 17:25:53 crc kubenswrapper[4777]: I1124 17:25:53.053886 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2rxw\" (UniqueName: \"kubernetes.io/projected/f27aca44-4a34-469c-97fd-075eed1300dd-kube-api-access-m2rxw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qfmq5\" (UID: \"f27aca44-4a34-469c-97fd-075eed1300dd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" Nov 24 17:25:53 crc kubenswrapper[4777]: I1124 17:25:53.054018 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f27aca44-4a34-469c-97fd-075eed1300dd-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qfmq5\" (UID: \"f27aca44-4a34-469c-97fd-075eed1300dd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" Nov 24 17:25:53 crc kubenswrapper[4777]: I1124 17:25:53.054056 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f27aca44-4a34-469c-97fd-075eed1300dd-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qfmq5\" (UID: \"f27aca44-4a34-469c-97fd-075eed1300dd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" Nov 24 17:25:53 crc kubenswrapper[4777]: I1124 17:25:53.058585 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f27aca44-4a34-469c-97fd-075eed1300dd-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qfmq5\" (UID: \"f27aca44-4a34-469c-97fd-075eed1300dd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" Nov 24 17:25:53 crc kubenswrapper[4777]: I1124 17:25:53.059028 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f27aca44-4a34-469c-97fd-075eed1300dd-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qfmq5\" (UID: \"f27aca44-4a34-469c-97fd-075eed1300dd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" Nov 24 17:25:53 crc kubenswrapper[4777]: I1124 17:25:53.075000 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2rxw\" (UniqueName: \"kubernetes.io/projected/f27aca44-4a34-469c-97fd-075eed1300dd-kube-api-access-m2rxw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qfmq5\" (UID: \"f27aca44-4a34-469c-97fd-075eed1300dd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" Nov 24 17:25:53 crc kubenswrapper[4777]: I1124 17:25:53.243278 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" Nov 24 17:25:53 crc kubenswrapper[4777]: I1124 17:25:53.796046 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-28c29" podUID="29d9b08e-8a4b-4e0b-9b83-724b303e2e61" containerName="registry-server" containerID="cri-o://1ad3184a6047d6ef934e6204aab80fb218594a62fdc8825dd10b23c6908aa3a7" gracePeriod=2 Nov 24 17:25:53 crc kubenswrapper[4777]: I1124 17:25:53.894668 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5"] Nov 24 17:25:53 crc kubenswrapper[4777]: W1124 17:25:53.975098 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf27aca44_4a34_469c_97fd_075eed1300dd.slice/crio-e6a28e9c8a04635f512c259afbb80008bd2c52c2ca5d8a0ce81f5cc232d02e8e WatchSource:0}: Error finding container e6a28e9c8a04635f512c259afbb80008bd2c52c2ca5d8a0ce81f5cc232d02e8e: Status 404 returned error can't find the container with id e6a28e9c8a04635f512c259afbb80008bd2c52c2ca5d8a0ce81f5cc232d02e8e Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.369623 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.386792 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-catalog-content\") pod \"29d9b08e-8a4b-4e0b-9b83-724b303e2e61\" (UID: \"29d9b08e-8a4b-4e0b-9b83-724b303e2e61\") " Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.388148 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5knll\" (UniqueName: \"kubernetes.io/projected/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-kube-api-access-5knll\") pod \"29d9b08e-8a4b-4e0b-9b83-724b303e2e61\" (UID: \"29d9b08e-8a4b-4e0b-9b83-724b303e2e61\") " Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.388426 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-utilities\") pod \"29d9b08e-8a4b-4e0b-9b83-724b303e2e61\" (UID: \"29d9b08e-8a4b-4e0b-9b83-724b303e2e61\") " Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.389027 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-utilities" (OuterVolumeSpecName: "utilities") pod "29d9b08e-8a4b-4e0b-9b83-724b303e2e61" (UID: "29d9b08e-8a4b-4e0b-9b83-724b303e2e61"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.389490 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.394285 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-kube-api-access-5knll" (OuterVolumeSpecName: "kube-api-access-5knll") pod "29d9b08e-8a4b-4e0b-9b83-724b303e2e61" (UID: "29d9b08e-8a4b-4e0b-9b83-724b303e2e61"). InnerVolumeSpecName "kube-api-access-5knll". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.422443 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "29d9b08e-8a4b-4e0b-9b83-724b303e2e61" (UID: "29d9b08e-8a4b-4e0b-9b83-724b303e2e61"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.490769 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.490798 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5knll\" (UniqueName: \"kubernetes.io/projected/29d9b08e-8a4b-4e0b-9b83-724b303e2e61-kube-api-access-5knll\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.813876 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" event={"ID":"f27aca44-4a34-469c-97fd-075eed1300dd","Type":"ContainerStarted","Data":"6359a906b293fb9b5443e03a9687c610046682db9dbb6e4aba983d499701c62a"} Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.814492 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" event={"ID":"f27aca44-4a34-469c-97fd-075eed1300dd","Type":"ContainerStarted","Data":"e6a28e9c8a04635f512c259afbb80008bd2c52c2ca5d8a0ce81f5cc232d02e8e"} Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.819505 4777 generic.go:334] "Generic (PLEG): container finished" podID="29d9b08e-8a4b-4e0b-9b83-724b303e2e61" containerID="1ad3184a6047d6ef934e6204aab80fb218594a62fdc8825dd10b23c6908aa3a7" exitCode=0 Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.819589 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28c29" event={"ID":"29d9b08e-8a4b-4e0b-9b83-724b303e2e61","Type":"ContainerDied","Data":"1ad3184a6047d6ef934e6204aab80fb218594a62fdc8825dd10b23c6908aa3a7"} Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.819636 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28c29" event={"ID":"29d9b08e-8a4b-4e0b-9b83-724b303e2e61","Type":"ContainerDied","Data":"799dd0e3391e59046adc8ee6250418d61567ee4bb40aa8029da6eaddd7547393"} Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.819662 4777 scope.go:117] "RemoveContainer" containerID="1ad3184a6047d6ef934e6204aab80fb218594a62fdc8825dd10b23c6908aa3a7" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.820094 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-28c29" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.852198 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" podStartSLOduration=2.390503364 podStartE2EDuration="2.852173365s" podCreationTimestamp="2025-11-24 17:25:52 +0000 UTC" firstStartedPulling="2025-11-24 17:25:54.024250185 +0000 UTC m=+1512.183445234" lastFinishedPulling="2025-11-24 17:25:54.485920186 +0000 UTC m=+1512.645115235" observedRunningTime="2025-11-24 17:25:54.83642371 +0000 UTC m=+1512.995618779" watchObservedRunningTime="2025-11-24 17:25:54.852173365 +0000 UTC m=+1513.011368424" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.869064 4777 scope.go:117] "RemoveContainer" containerID="ad0dc0bb5c166a1aaab711bd7499b7809107815948f8a5b52a2c335d91c04768" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.877094 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-28c29"] Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.893456 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-28c29"] Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.909143 4777 scope.go:117] "RemoveContainer" containerID="a342a44e213990dd559f9ae6642b7599b942cd3d74840f61260e89ed4acfb4c1" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.936560 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-np6rr" podUID="a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" containerName="registry-server" probeResult="failure" output=< Nov 24 17:25:54 crc kubenswrapper[4777]: timeout: failed to connect service ":50051" within 1s Nov 24 17:25:54 crc kubenswrapper[4777]: > Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.936859 4777 scope.go:117] "RemoveContainer" containerID="1ad3184a6047d6ef934e6204aab80fb218594a62fdc8825dd10b23c6908aa3a7" Nov 24 17:25:54 crc kubenswrapper[4777]: E1124 17:25:54.937297 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ad3184a6047d6ef934e6204aab80fb218594a62fdc8825dd10b23c6908aa3a7\": container with ID starting with 1ad3184a6047d6ef934e6204aab80fb218594a62fdc8825dd10b23c6908aa3a7 not found: ID does not exist" containerID="1ad3184a6047d6ef934e6204aab80fb218594a62fdc8825dd10b23c6908aa3a7" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.937327 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ad3184a6047d6ef934e6204aab80fb218594a62fdc8825dd10b23c6908aa3a7"} err="failed to get container status \"1ad3184a6047d6ef934e6204aab80fb218594a62fdc8825dd10b23c6908aa3a7\": rpc error: code = NotFound desc = could not find container \"1ad3184a6047d6ef934e6204aab80fb218594a62fdc8825dd10b23c6908aa3a7\": container with ID starting with 1ad3184a6047d6ef934e6204aab80fb218594a62fdc8825dd10b23c6908aa3a7 not found: ID does not exist" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.937350 4777 scope.go:117] "RemoveContainer" containerID="ad0dc0bb5c166a1aaab711bd7499b7809107815948f8a5b52a2c335d91c04768" Nov 24 17:25:54 crc kubenswrapper[4777]: E1124 17:25:54.937679 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad0dc0bb5c166a1aaab711bd7499b7809107815948f8a5b52a2c335d91c04768\": container with ID starting with ad0dc0bb5c166a1aaab711bd7499b7809107815948f8a5b52a2c335d91c04768 not found: ID does not exist" containerID="ad0dc0bb5c166a1aaab711bd7499b7809107815948f8a5b52a2c335d91c04768" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.937703 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad0dc0bb5c166a1aaab711bd7499b7809107815948f8a5b52a2c335d91c04768"} err="failed to get container status \"ad0dc0bb5c166a1aaab711bd7499b7809107815948f8a5b52a2c335d91c04768\": rpc error: code = NotFound desc = could not find container \"ad0dc0bb5c166a1aaab711bd7499b7809107815948f8a5b52a2c335d91c04768\": container with ID starting with ad0dc0bb5c166a1aaab711bd7499b7809107815948f8a5b52a2c335d91c04768 not found: ID does not exist" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.937722 4777 scope.go:117] "RemoveContainer" containerID="a342a44e213990dd559f9ae6642b7599b942cd3d74840f61260e89ed4acfb4c1" Nov 24 17:25:54 crc kubenswrapper[4777]: E1124 17:25:54.938115 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a342a44e213990dd559f9ae6642b7599b942cd3d74840f61260e89ed4acfb4c1\": container with ID starting with a342a44e213990dd559f9ae6642b7599b942cd3d74840f61260e89ed4acfb4c1 not found: ID does not exist" containerID="a342a44e213990dd559f9ae6642b7599b942cd3d74840f61260e89ed4acfb4c1" Nov 24 17:25:54 crc kubenswrapper[4777]: I1124 17:25:54.938137 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a342a44e213990dd559f9ae6642b7599b942cd3d74840f61260e89ed4acfb4c1"} err="failed to get container status \"a342a44e213990dd559f9ae6642b7599b942cd3d74840f61260e89ed4acfb4c1\": rpc error: code = NotFound desc = could not find container \"a342a44e213990dd559f9ae6642b7599b942cd3d74840f61260e89ed4acfb4c1\": container with ID starting with a342a44e213990dd559f9ae6642b7599b942cd3d74840f61260e89ed4acfb4c1 not found: ID does not exist" Nov 24 17:25:55 crc kubenswrapper[4777]: I1124 17:25:55.262147 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29d9b08e-8a4b-4e0b-9b83-724b303e2e61" path="/var/lib/kubelet/pods/29d9b08e-8a4b-4e0b-9b83-724b303e2e61/volumes" Nov 24 17:25:55 crc kubenswrapper[4777]: I1124 17:25:55.876278 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 17:25:57 crc kubenswrapper[4777]: I1124 17:25:57.858419 4777 generic.go:334] "Generic (PLEG): container finished" podID="f27aca44-4a34-469c-97fd-075eed1300dd" containerID="6359a906b293fb9b5443e03a9687c610046682db9dbb6e4aba983d499701c62a" exitCode=0 Nov 24 17:25:57 crc kubenswrapper[4777]: I1124 17:25:57.858478 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" event={"ID":"f27aca44-4a34-469c-97fd-075eed1300dd","Type":"ContainerDied","Data":"6359a906b293fb9b5443e03a9687c610046682db9dbb6e4aba983d499701c62a"} Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.464930 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.616241 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2rxw\" (UniqueName: \"kubernetes.io/projected/f27aca44-4a34-469c-97fd-075eed1300dd-kube-api-access-m2rxw\") pod \"f27aca44-4a34-469c-97fd-075eed1300dd\" (UID: \"f27aca44-4a34-469c-97fd-075eed1300dd\") " Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.616367 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f27aca44-4a34-469c-97fd-075eed1300dd-inventory\") pod \"f27aca44-4a34-469c-97fd-075eed1300dd\" (UID: \"f27aca44-4a34-469c-97fd-075eed1300dd\") " Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.616419 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f27aca44-4a34-469c-97fd-075eed1300dd-ssh-key\") pod \"f27aca44-4a34-469c-97fd-075eed1300dd\" (UID: \"f27aca44-4a34-469c-97fd-075eed1300dd\") " Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.626231 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f27aca44-4a34-469c-97fd-075eed1300dd-kube-api-access-m2rxw" (OuterVolumeSpecName: "kube-api-access-m2rxw") pod "f27aca44-4a34-469c-97fd-075eed1300dd" (UID: "f27aca44-4a34-469c-97fd-075eed1300dd"). InnerVolumeSpecName "kube-api-access-m2rxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.678851 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f27aca44-4a34-469c-97fd-075eed1300dd-inventory" (OuterVolumeSpecName: "inventory") pod "f27aca44-4a34-469c-97fd-075eed1300dd" (UID: "f27aca44-4a34-469c-97fd-075eed1300dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.681158 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f27aca44-4a34-469c-97fd-075eed1300dd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f27aca44-4a34-469c-97fd-075eed1300dd" (UID: "f27aca44-4a34-469c-97fd-075eed1300dd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.719025 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2rxw\" (UniqueName: \"kubernetes.io/projected/f27aca44-4a34-469c-97fd-075eed1300dd-kube-api-access-m2rxw\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.719067 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f27aca44-4a34-469c-97fd-075eed1300dd-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.719079 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f27aca44-4a34-469c-97fd-075eed1300dd-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.885932 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" event={"ID":"f27aca44-4a34-469c-97fd-075eed1300dd","Type":"ContainerDied","Data":"e6a28e9c8a04635f512c259afbb80008bd2c52c2ca5d8a0ce81f5cc232d02e8e"} Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.886152 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6a28e9c8a04635f512c259afbb80008bd2c52c2ca5d8a0ce81f5cc232d02e8e" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.886007 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qfmq5" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.971264 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g"] Nov 24 17:25:59 crc kubenswrapper[4777]: E1124 17:25:59.972041 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29d9b08e-8a4b-4e0b-9b83-724b303e2e61" containerName="extract-utilities" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.972108 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="29d9b08e-8a4b-4e0b-9b83-724b303e2e61" containerName="extract-utilities" Nov 24 17:25:59 crc kubenswrapper[4777]: E1124 17:25:59.972204 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29d9b08e-8a4b-4e0b-9b83-724b303e2e61" containerName="extract-content" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.972257 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="29d9b08e-8a4b-4e0b-9b83-724b303e2e61" containerName="extract-content" Nov 24 17:25:59 crc kubenswrapper[4777]: E1124 17:25:59.972315 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29d9b08e-8a4b-4e0b-9b83-724b303e2e61" containerName="registry-server" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.972368 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="29d9b08e-8a4b-4e0b-9b83-724b303e2e61" containerName="registry-server" Nov 24 17:25:59 crc kubenswrapper[4777]: E1124 17:25:59.972430 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f27aca44-4a34-469c-97fd-075eed1300dd" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.972478 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="f27aca44-4a34-469c-97fd-075eed1300dd" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.972721 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="f27aca44-4a34-469c-97fd-075eed1300dd" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.972789 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="29d9b08e-8a4b-4e0b-9b83-724b303e2e61" containerName="registry-server" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.973617 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:25:59 crc kubenswrapper[4777]: I1124 17:25:59.983750 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g"] Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.010530 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.010905 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.010933 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.015631 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.027592 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.027678 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9jn9\" (UniqueName: \"kubernetes.io/projected/1bda09cc-b84e-409f-b4cd-fc387aa02c61-kube-api-access-t9jn9\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.027703 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.027762 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.129099 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9jn9\" (UniqueName: \"kubernetes.io/projected/1bda09cc-b84e-409f-b4cd-fc387aa02c61-kube-api-access-t9jn9\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.129464 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.129530 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.129627 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.137093 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.137198 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.138138 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.150287 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9jn9\" (UniqueName: \"kubernetes.io/projected/1bda09cc-b84e-409f-b4cd-fc387aa02c61-kube-api-access-t9jn9\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.329694 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:26:00 crc kubenswrapper[4777]: I1124 17:26:00.902823 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g"] Nov 24 17:26:00 crc kubenswrapper[4777]: W1124 17:26:00.905133 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bda09cc_b84e_409f_b4cd_fc387aa02c61.slice/crio-1d3e347a732736f96f45bb23b3357a632ccd4b1788bac7e984cb89ff3e367470 WatchSource:0}: Error finding container 1d3e347a732736f96f45bb23b3357a632ccd4b1788bac7e984cb89ff3e367470: Status 404 returned error can't find the container with id 1d3e347a732736f96f45bb23b3357a632ccd4b1788bac7e984cb89ff3e367470 Nov 24 17:26:01 crc kubenswrapper[4777]: I1124 17:26:01.924232 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" event={"ID":"1bda09cc-b84e-409f-b4cd-fc387aa02c61","Type":"ContainerStarted","Data":"dc6dcab6cad86bd8132f052bfc96b5dbf36e23b4ede7a282ceb8d4a2bbbb31d3"} Nov 24 17:26:01 crc kubenswrapper[4777]: I1124 17:26:01.924577 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" event={"ID":"1bda09cc-b84e-409f-b4cd-fc387aa02c61","Type":"ContainerStarted","Data":"1d3e347a732736f96f45bb23b3357a632ccd4b1788bac7e984cb89ff3e367470"} Nov 24 17:26:01 crc kubenswrapper[4777]: I1124 17:26:01.952286 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" podStartSLOduration=2.506573958 podStartE2EDuration="2.952269656s" podCreationTimestamp="2025-11-24 17:25:59 +0000 UTC" firstStartedPulling="2025-11-24 17:26:00.90785496 +0000 UTC m=+1519.067050009" lastFinishedPulling="2025-11-24 17:26:01.353550648 +0000 UTC m=+1519.512745707" observedRunningTime="2025-11-24 17:26:01.943332407 +0000 UTC m=+1520.102527456" watchObservedRunningTime="2025-11-24 17:26:01.952269656 +0000 UTC m=+1520.111464705" Nov 24 17:26:03 crc kubenswrapper[4777]: I1124 17:26:03.957472 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:26:04 crc kubenswrapper[4777]: I1124 17:26:04.029739 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:26:04 crc kubenswrapper[4777]: I1124 17:26:04.747583 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-np6rr"] Nov 24 17:26:05 crc kubenswrapper[4777]: I1124 17:26:05.959235 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-np6rr" podUID="a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" containerName="registry-server" containerID="cri-o://892bffab92874cd3da4b34c52cc1f57b8f17d8db9e79e13dc1cc58ee82d43d5f" gracePeriod=2 Nov 24 17:26:06 crc kubenswrapper[4777]: I1124 17:26:06.505920 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:26:06 crc kubenswrapper[4777]: I1124 17:26:06.664518 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2w7g\" (UniqueName: \"kubernetes.io/projected/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-kube-api-access-g2w7g\") pod \"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a\" (UID: \"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a\") " Nov 24 17:26:06 crc kubenswrapper[4777]: I1124 17:26:06.664577 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-catalog-content\") pod \"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a\" (UID: \"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a\") " Nov 24 17:26:06 crc kubenswrapper[4777]: I1124 17:26:06.664744 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-utilities\") pod \"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a\" (UID: \"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a\") " Nov 24 17:26:06 crc kubenswrapper[4777]: I1124 17:26:06.665929 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-utilities" (OuterVolumeSpecName: "utilities") pod "a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" (UID: "a5ec65c8-b4cd-4917-bdc0-533df6b58c0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:26:06 crc kubenswrapper[4777]: I1124 17:26:06.673642 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-kube-api-access-g2w7g" (OuterVolumeSpecName: "kube-api-access-g2w7g") pod "a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" (UID: "a5ec65c8-b4cd-4917-bdc0-533df6b58c0a"). InnerVolumeSpecName "kube-api-access-g2w7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:26:06 crc kubenswrapper[4777]: I1124 17:26:06.728504 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" (UID: "a5ec65c8-b4cd-4917-bdc0-533df6b58c0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:26:06 crc kubenswrapper[4777]: I1124 17:26:06.766704 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2w7g\" (UniqueName: \"kubernetes.io/projected/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-kube-api-access-g2w7g\") on node \"crc\" DevicePath \"\"" Nov 24 17:26:06 crc kubenswrapper[4777]: I1124 17:26:06.766734 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:26:06 crc kubenswrapper[4777]: I1124 17:26:06.766743 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:26:06 crc kubenswrapper[4777]: I1124 17:26:06.971253 4777 generic.go:334] "Generic (PLEG): container finished" podID="a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" containerID="892bffab92874cd3da4b34c52cc1f57b8f17d8db9e79e13dc1cc58ee82d43d5f" exitCode=0 Nov 24 17:26:06 crc kubenswrapper[4777]: I1124 17:26:06.971610 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-np6rr" event={"ID":"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a","Type":"ContainerDied","Data":"892bffab92874cd3da4b34c52cc1f57b8f17d8db9e79e13dc1cc58ee82d43d5f"} Nov 24 17:26:06 crc kubenswrapper[4777]: I1124 17:26:06.971643 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-np6rr" event={"ID":"a5ec65c8-b4cd-4917-bdc0-533df6b58c0a","Type":"ContainerDied","Data":"be4ed43cf96273ccd81fa07484b867aeab3248b7d856fc0669198f45e9b25592"} Nov 24 17:26:06 crc kubenswrapper[4777]: I1124 17:26:06.971664 4777 scope.go:117] "RemoveContainer" containerID="892bffab92874cd3da4b34c52cc1f57b8f17d8db9e79e13dc1cc58ee82d43d5f" Nov 24 17:26:06 crc kubenswrapper[4777]: I1124 17:26:06.971807 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-np6rr" Nov 24 17:26:07 crc kubenswrapper[4777]: I1124 17:26:07.012949 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-np6rr"] Nov 24 17:26:07 crc kubenswrapper[4777]: I1124 17:26:07.013066 4777 scope.go:117] "RemoveContainer" containerID="b477dfa57e59c709c249d2391554690f6de794a353d2763eb31600c3a316397b" Nov 24 17:26:07 crc kubenswrapper[4777]: I1124 17:26:07.030437 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-np6rr"] Nov 24 17:26:07 crc kubenswrapper[4777]: I1124 17:26:07.036908 4777 scope.go:117] "RemoveContainer" containerID="fdd99f8b94c8650213b66b985a0fc99847836973cf078e720ba66c0fad66ced4" Nov 24 17:26:07 crc kubenswrapper[4777]: I1124 17:26:07.110239 4777 scope.go:117] "RemoveContainer" containerID="892bffab92874cd3da4b34c52cc1f57b8f17d8db9e79e13dc1cc58ee82d43d5f" Nov 24 17:26:07 crc kubenswrapper[4777]: E1124 17:26:07.110717 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"892bffab92874cd3da4b34c52cc1f57b8f17d8db9e79e13dc1cc58ee82d43d5f\": container with ID starting with 892bffab92874cd3da4b34c52cc1f57b8f17d8db9e79e13dc1cc58ee82d43d5f not found: ID does not exist" containerID="892bffab92874cd3da4b34c52cc1f57b8f17d8db9e79e13dc1cc58ee82d43d5f" Nov 24 17:26:07 crc kubenswrapper[4777]: I1124 17:26:07.110746 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"892bffab92874cd3da4b34c52cc1f57b8f17d8db9e79e13dc1cc58ee82d43d5f"} err="failed to get container status \"892bffab92874cd3da4b34c52cc1f57b8f17d8db9e79e13dc1cc58ee82d43d5f\": rpc error: code = NotFound desc = could not find container \"892bffab92874cd3da4b34c52cc1f57b8f17d8db9e79e13dc1cc58ee82d43d5f\": container with ID starting with 892bffab92874cd3da4b34c52cc1f57b8f17d8db9e79e13dc1cc58ee82d43d5f not found: ID does not exist" Nov 24 17:26:07 crc kubenswrapper[4777]: I1124 17:26:07.110767 4777 scope.go:117] "RemoveContainer" containerID="b477dfa57e59c709c249d2391554690f6de794a353d2763eb31600c3a316397b" Nov 24 17:26:07 crc kubenswrapper[4777]: E1124 17:26:07.111231 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b477dfa57e59c709c249d2391554690f6de794a353d2763eb31600c3a316397b\": container with ID starting with b477dfa57e59c709c249d2391554690f6de794a353d2763eb31600c3a316397b not found: ID does not exist" containerID="b477dfa57e59c709c249d2391554690f6de794a353d2763eb31600c3a316397b" Nov 24 17:26:07 crc kubenswrapper[4777]: I1124 17:26:07.111259 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b477dfa57e59c709c249d2391554690f6de794a353d2763eb31600c3a316397b"} err="failed to get container status \"b477dfa57e59c709c249d2391554690f6de794a353d2763eb31600c3a316397b\": rpc error: code = NotFound desc = could not find container \"b477dfa57e59c709c249d2391554690f6de794a353d2763eb31600c3a316397b\": container with ID starting with b477dfa57e59c709c249d2391554690f6de794a353d2763eb31600c3a316397b not found: ID does not exist" Nov 24 17:26:07 crc kubenswrapper[4777]: I1124 17:26:07.111274 4777 scope.go:117] "RemoveContainer" containerID="fdd99f8b94c8650213b66b985a0fc99847836973cf078e720ba66c0fad66ced4" Nov 24 17:26:07 crc kubenswrapper[4777]: E1124 17:26:07.111553 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdd99f8b94c8650213b66b985a0fc99847836973cf078e720ba66c0fad66ced4\": container with ID starting with fdd99f8b94c8650213b66b985a0fc99847836973cf078e720ba66c0fad66ced4 not found: ID does not exist" containerID="fdd99f8b94c8650213b66b985a0fc99847836973cf078e720ba66c0fad66ced4" Nov 24 17:26:07 crc kubenswrapper[4777]: I1124 17:26:07.111575 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdd99f8b94c8650213b66b985a0fc99847836973cf078e720ba66c0fad66ced4"} err="failed to get container status \"fdd99f8b94c8650213b66b985a0fc99847836973cf078e720ba66c0fad66ced4\": rpc error: code = NotFound desc = could not find container \"fdd99f8b94c8650213b66b985a0fc99847836973cf078e720ba66c0fad66ced4\": container with ID starting with fdd99f8b94c8650213b66b985a0fc99847836973cf078e720ba66c0fad66ced4 not found: ID does not exist" Nov 24 17:26:07 crc kubenswrapper[4777]: I1124 17:26:07.258360 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" path="/var/lib/kubelet/pods/a5ec65c8-b4cd-4917-bdc0-533df6b58c0a/volumes" Nov 24 17:26:14 crc kubenswrapper[4777]: I1124 17:26:14.272704 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:26:14 crc kubenswrapper[4777]: I1124 17:26:14.273468 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:26:14 crc kubenswrapper[4777]: I1124 17:26:14.273531 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:26:14 crc kubenswrapper[4777]: I1124 17:26:14.274736 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:26:14 crc kubenswrapper[4777]: I1124 17:26:14.274855 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" gracePeriod=600 Nov 24 17:26:14 crc kubenswrapper[4777]: E1124 17:26:14.403994 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:26:15 crc kubenswrapper[4777]: I1124 17:26:15.069216 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" exitCode=0 Nov 24 17:26:15 crc kubenswrapper[4777]: I1124 17:26:15.069274 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4"} Nov 24 17:26:15 crc kubenswrapper[4777]: I1124 17:26:15.069328 4777 scope.go:117] "RemoveContainer" containerID="a6faf411f1a27fb01d7bc7ffb8b05ff8c95006e6030cf4cec387d5e3b99354f9" Nov 24 17:26:15 crc kubenswrapper[4777]: I1124 17:26:15.069888 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:26:15 crc kubenswrapper[4777]: E1124 17:26:15.070180 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:26:27 crc kubenswrapper[4777]: I1124 17:26:27.245808 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:26:27 crc kubenswrapper[4777]: E1124 17:26:27.246889 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:26:42 crc kubenswrapper[4777]: I1124 17:26:42.245378 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:26:42 crc kubenswrapper[4777]: E1124 17:26:42.246162 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:26:45 crc kubenswrapper[4777]: I1124 17:26:45.539517 4777 scope.go:117] "RemoveContainer" containerID="ce26c2638bc8a5a44ac408f9d18326426457cc38ade16ed34b06be16814f2ac4" Nov 24 17:26:45 crc kubenswrapper[4777]: I1124 17:26:45.597911 4777 scope.go:117] "RemoveContainer" containerID="1c6e63bcd736b4e0d34d704190b35d5a66b8cc042af83e7c9a57888c102700a2" Nov 24 17:26:55 crc kubenswrapper[4777]: I1124 17:26:55.246055 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:26:55 crc kubenswrapper[4777]: E1124 17:26:55.246991 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:27:06 crc kubenswrapper[4777]: I1124 17:27:06.245774 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:27:06 crc kubenswrapper[4777]: E1124 17:27:06.246680 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:27:17 crc kubenswrapper[4777]: I1124 17:27:17.245714 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:27:17 crc kubenswrapper[4777]: E1124 17:27:17.247939 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:27:28 crc kubenswrapper[4777]: I1124 17:27:28.245817 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:27:28 crc kubenswrapper[4777]: E1124 17:27:28.246942 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:27:42 crc kubenswrapper[4777]: I1124 17:27:42.245255 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:27:42 crc kubenswrapper[4777]: E1124 17:27:42.246888 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:27:45 crc kubenswrapper[4777]: I1124 17:27:45.748294 4777 scope.go:117] "RemoveContainer" containerID="8ec747f5df55dc39b3d1b3cf4cc66787156bf54495e2daaed8ebb86d86f29b99" Nov 24 17:27:45 crc kubenswrapper[4777]: I1124 17:27:45.783613 4777 scope.go:117] "RemoveContainer" containerID="b49906183d185d99ed7bf30b881efa446b8a6139d8008b6ee5e4218e587e7324" Nov 24 17:27:45 crc kubenswrapper[4777]: I1124 17:27:45.879085 4777 scope.go:117] "RemoveContainer" containerID="efdb2752b15ceedc4b1933eabf81a88f9b4c0110a61f8e05eb3f4dbe01cd537c" Nov 24 17:27:45 crc kubenswrapper[4777]: I1124 17:27:45.913435 4777 scope.go:117] "RemoveContainer" containerID="80a456e5941ed70510ff0ae6327a8bc1cea6d791e29114eeee24f07063cf8411" Nov 24 17:27:45 crc kubenswrapper[4777]: I1124 17:27:45.936342 4777 scope.go:117] "RemoveContainer" containerID="8533cbb4cbc2c29d634a0c93caeff0ed22f0bdb3caca2ebb76c693d6aa9aec1e" Nov 24 17:27:53 crc kubenswrapper[4777]: I1124 17:27:53.252088 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:27:53 crc kubenswrapper[4777]: E1124 17:27:53.253200 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:28:08 crc kubenswrapper[4777]: I1124 17:28:08.246320 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:28:08 crc kubenswrapper[4777]: E1124 17:28:08.247570 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:28:20 crc kubenswrapper[4777]: I1124 17:28:20.245858 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:28:20 crc kubenswrapper[4777]: E1124 17:28:20.246598 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:28:33 crc kubenswrapper[4777]: I1124 17:28:33.262384 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:28:33 crc kubenswrapper[4777]: E1124 17:28:33.263199 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:28:45 crc kubenswrapper[4777]: I1124 17:28:45.245407 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:28:45 crc kubenswrapper[4777]: E1124 17:28:45.246124 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:28:46 crc kubenswrapper[4777]: I1124 17:28:46.022011 4777 scope.go:117] "RemoveContainer" containerID="fa36391d822c8dd41c854944890ec302f52245eb3aa7a319184bd147967a8c9d" Nov 24 17:28:57 crc kubenswrapper[4777]: I1124 17:28:57.245583 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:28:57 crc kubenswrapper[4777]: E1124 17:28:57.246350 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:29:12 crc kubenswrapper[4777]: I1124 17:29:12.245760 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:29:12 crc kubenswrapper[4777]: E1124 17:29:12.247206 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:29:13 crc kubenswrapper[4777]: I1124 17:29:13.307210 4777 generic.go:334] "Generic (PLEG): container finished" podID="1bda09cc-b84e-409f-b4cd-fc387aa02c61" containerID="dc6dcab6cad86bd8132f052bfc96b5dbf36e23b4ede7a282ceb8d4a2bbbb31d3" exitCode=0 Nov 24 17:29:13 crc kubenswrapper[4777]: I1124 17:29:13.307273 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" event={"ID":"1bda09cc-b84e-409f-b4cd-fc387aa02c61","Type":"ContainerDied","Data":"dc6dcab6cad86bd8132f052bfc96b5dbf36e23b4ede7a282ceb8d4a2bbbb31d3"} Nov 24 17:29:14 crc kubenswrapper[4777]: I1124 17:29:14.889194 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:29:14 crc kubenswrapper[4777]: I1124 17:29:14.988189 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-inventory\") pod \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " Nov 24 17:29:14 crc kubenswrapper[4777]: I1124 17:29:14.988334 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-ssh-key\") pod \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " Nov 24 17:29:14 crc kubenswrapper[4777]: I1124 17:29:14.988402 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9jn9\" (UniqueName: \"kubernetes.io/projected/1bda09cc-b84e-409f-b4cd-fc387aa02c61-kube-api-access-t9jn9\") pod \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " Nov 24 17:29:14 crc kubenswrapper[4777]: I1124 17:29:14.988653 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-bootstrap-combined-ca-bundle\") pod \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\" (UID: \"1bda09cc-b84e-409f-b4cd-fc387aa02c61\") " Nov 24 17:29:14 crc kubenswrapper[4777]: I1124 17:29:14.999334 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bda09cc-b84e-409f-b4cd-fc387aa02c61-kube-api-access-t9jn9" (OuterVolumeSpecName: "kube-api-access-t9jn9") pod "1bda09cc-b84e-409f-b4cd-fc387aa02c61" (UID: "1bda09cc-b84e-409f-b4cd-fc387aa02c61"). InnerVolumeSpecName "kube-api-access-t9jn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:29:14 crc kubenswrapper[4777]: I1124 17:29:14.999835 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "1bda09cc-b84e-409f-b4cd-fc387aa02c61" (UID: "1bda09cc-b84e-409f-b4cd-fc387aa02c61"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.040340 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1bda09cc-b84e-409f-b4cd-fc387aa02c61" (UID: "1bda09cc-b84e-409f-b4cd-fc387aa02c61"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.040837 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-inventory" (OuterVolumeSpecName: "inventory") pod "1bda09cc-b84e-409f-b4cd-fc387aa02c61" (UID: "1bda09cc-b84e-409f-b4cd-fc387aa02c61"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.092141 4777 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.092234 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.092245 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1bda09cc-b84e-409f-b4cd-fc387aa02c61-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.092279 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9jn9\" (UniqueName: \"kubernetes.io/projected/1bda09cc-b84e-409f-b4cd-fc387aa02c61-kube-api-access-t9jn9\") on node \"crc\" DevicePath \"\"" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.331525 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" event={"ID":"1bda09cc-b84e-409f-b4cd-fc387aa02c61","Type":"ContainerDied","Data":"1d3e347a732736f96f45bb23b3357a632ccd4b1788bac7e984cb89ff3e367470"} Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.331566 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.331581 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d3e347a732736f96f45bb23b3357a632ccd4b1788bac7e984cb89ff3e367470" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.415429 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j"] Nov 24 17:29:15 crc kubenswrapper[4777]: E1124 17:29:15.415831 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" containerName="extract-utilities" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.415846 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" containerName="extract-utilities" Nov 24 17:29:15 crc kubenswrapper[4777]: E1124 17:29:15.415867 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" containerName="extract-content" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.415876 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" containerName="extract-content" Nov 24 17:29:15 crc kubenswrapper[4777]: E1124 17:29:15.415897 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" containerName="registry-server" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.415903 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" containerName="registry-server" Nov 24 17:29:15 crc kubenswrapper[4777]: E1124 17:29:15.415917 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bda09cc-b84e-409f-b4cd-fc387aa02c61" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.415924 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bda09cc-b84e-409f-b4cd-fc387aa02c61" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.416120 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5ec65c8-b4cd-4917-bdc0-533df6b58c0a" containerName="registry-server" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.416150 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bda09cc-b84e-409f-b4cd-fc387aa02c61" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.416825 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.419286 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.419298 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.419496 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.419616 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.439029 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j"] Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.500276 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/266eceac-bfd7-4597-b218-62cea462ce5c-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j\" (UID: \"266eceac-bfd7-4597-b218-62cea462ce5c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.500584 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/266eceac-bfd7-4597-b218-62cea462ce5c-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j\" (UID: \"266eceac-bfd7-4597-b218-62cea462ce5c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.500763 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntgrj\" (UniqueName: \"kubernetes.io/projected/266eceac-bfd7-4597-b218-62cea462ce5c-kube-api-access-ntgrj\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j\" (UID: \"266eceac-bfd7-4597-b218-62cea462ce5c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.602853 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/266eceac-bfd7-4597-b218-62cea462ce5c-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j\" (UID: \"266eceac-bfd7-4597-b218-62cea462ce5c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.602991 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/266eceac-bfd7-4597-b218-62cea462ce5c-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j\" (UID: \"266eceac-bfd7-4597-b218-62cea462ce5c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.603037 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntgrj\" (UniqueName: \"kubernetes.io/projected/266eceac-bfd7-4597-b218-62cea462ce5c-kube-api-access-ntgrj\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j\" (UID: \"266eceac-bfd7-4597-b218-62cea462ce5c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.607135 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/266eceac-bfd7-4597-b218-62cea462ce5c-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j\" (UID: \"266eceac-bfd7-4597-b218-62cea462ce5c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.607187 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/266eceac-bfd7-4597-b218-62cea462ce5c-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j\" (UID: \"266eceac-bfd7-4597-b218-62cea462ce5c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.639633 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntgrj\" (UniqueName: \"kubernetes.io/projected/266eceac-bfd7-4597-b218-62cea462ce5c-kube-api-access-ntgrj\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j\" (UID: \"266eceac-bfd7-4597-b218-62cea462ce5c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" Nov 24 17:29:15 crc kubenswrapper[4777]: I1124 17:29:15.740457 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" Nov 24 17:29:16 crc kubenswrapper[4777]: I1124 17:29:16.355119 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j"] Nov 24 17:29:17 crc kubenswrapper[4777]: I1124 17:29:17.351603 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" event={"ID":"266eceac-bfd7-4597-b218-62cea462ce5c","Type":"ContainerStarted","Data":"7580d122f9bad9d354197b2b9370d1a92a248cd12f7d0d09e9c1b3afcec029d4"} Nov 24 17:29:17 crc kubenswrapper[4777]: I1124 17:29:17.351654 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" event={"ID":"266eceac-bfd7-4597-b218-62cea462ce5c","Type":"ContainerStarted","Data":"aff38be9e6432737eecd36e8cf41e3aafa02764783da9ea0216eaf1cb85dd6a3"} Nov 24 17:29:17 crc kubenswrapper[4777]: I1124 17:29:17.389888 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" podStartSLOduration=1.955850223 podStartE2EDuration="2.389843309s" podCreationTimestamp="2025-11-24 17:29:15 +0000 UTC" firstStartedPulling="2025-11-24 17:29:16.352212653 +0000 UTC m=+1714.511407712" lastFinishedPulling="2025-11-24 17:29:16.786205749 +0000 UTC m=+1714.945400798" observedRunningTime="2025-11-24 17:29:17.37376143 +0000 UTC m=+1715.532956489" watchObservedRunningTime="2025-11-24 17:29:17.389843309 +0000 UTC m=+1715.549038358" Nov 24 17:29:23 crc kubenswrapper[4777]: I1124 17:29:23.258284 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:29:23 crc kubenswrapper[4777]: E1124 17:29:23.260722 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:29:38 crc kubenswrapper[4777]: I1124 17:29:38.245659 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:29:38 crc kubenswrapper[4777]: E1124 17:29:38.246571 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:29:46 crc kubenswrapper[4777]: I1124 17:29:46.101824 4777 scope.go:117] "RemoveContainer" containerID="f58ae0ff9d205c07f2d8d7a420739188b36a3777636f74796eec5753453b8abd" Nov 24 17:29:49 crc kubenswrapper[4777]: I1124 17:29:49.246164 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:29:49 crc kubenswrapper[4777]: E1124 17:29:49.246792 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.161701 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m"] Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.163816 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.166269 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.167019 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.179238 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m"] Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.299086 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fbbf0f0-b4c2-4e65-be72-d50116119445-secret-volume\") pod \"collect-profiles-29400090-vwx4m\" (UID: \"5fbbf0f0-b4c2-4e65-be72-d50116119445\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.299192 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fbbf0f0-b4c2-4e65-be72-d50116119445-config-volume\") pod \"collect-profiles-29400090-vwx4m\" (UID: \"5fbbf0f0-b4c2-4e65-be72-d50116119445\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.299243 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwf5x\" (UniqueName: \"kubernetes.io/projected/5fbbf0f0-b4c2-4e65-be72-d50116119445-kube-api-access-wwf5x\") pod \"collect-profiles-29400090-vwx4m\" (UID: \"5fbbf0f0-b4c2-4e65-be72-d50116119445\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.401489 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fbbf0f0-b4c2-4e65-be72-d50116119445-secret-volume\") pod \"collect-profiles-29400090-vwx4m\" (UID: \"5fbbf0f0-b4c2-4e65-be72-d50116119445\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.401594 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fbbf0f0-b4c2-4e65-be72-d50116119445-config-volume\") pod \"collect-profiles-29400090-vwx4m\" (UID: \"5fbbf0f0-b4c2-4e65-be72-d50116119445\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.401646 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwf5x\" (UniqueName: \"kubernetes.io/projected/5fbbf0f0-b4c2-4e65-be72-d50116119445-kube-api-access-wwf5x\") pod \"collect-profiles-29400090-vwx4m\" (UID: \"5fbbf0f0-b4c2-4e65-be72-d50116119445\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.402471 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fbbf0f0-b4c2-4e65-be72-d50116119445-config-volume\") pod \"collect-profiles-29400090-vwx4m\" (UID: \"5fbbf0f0-b4c2-4e65-be72-d50116119445\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.421769 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fbbf0f0-b4c2-4e65-be72-d50116119445-secret-volume\") pod \"collect-profiles-29400090-vwx4m\" (UID: \"5fbbf0f0-b4c2-4e65-be72-d50116119445\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.433923 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwf5x\" (UniqueName: \"kubernetes.io/projected/5fbbf0f0-b4c2-4e65-be72-d50116119445-kube-api-access-wwf5x\") pod \"collect-profiles-29400090-vwx4m\" (UID: \"5fbbf0f0-b4c2-4e65-be72-d50116119445\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.497704 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" Nov 24 17:30:00 crc kubenswrapper[4777]: I1124 17:30:00.975410 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m"] Nov 24 17:30:01 crc kubenswrapper[4777]: I1124 17:30:01.047936 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-mxkvt"] Nov 24 17:30:01 crc kubenswrapper[4777]: I1124 17:30:01.061391 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-98bd-account-create-x5c52"] Nov 24 17:30:01 crc kubenswrapper[4777]: I1124 17:30:01.070830 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-0e91-account-create-jvbdx"] Nov 24 17:30:01 crc kubenswrapper[4777]: I1124 17:30:01.079955 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-mxkvt"] Nov 24 17:30:01 crc kubenswrapper[4777]: I1124 17:30:01.089576 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-98bd-account-create-x5c52"] Nov 24 17:30:01 crc kubenswrapper[4777]: I1124 17:30:01.098609 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-0e91-account-create-jvbdx"] Nov 24 17:30:01 crc kubenswrapper[4777]: I1124 17:30:01.262483 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="947652ce-1c76-4d0e-9234-212063ae51c7" path="/var/lib/kubelet/pods/947652ce-1c76-4d0e-9234-212063ae51c7/volumes" Nov 24 17:30:01 crc kubenswrapper[4777]: I1124 17:30:01.263726 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="995fd0aa-138b-47b4-a001-3098fcf14e08" path="/var/lib/kubelet/pods/995fd0aa-138b-47b4-a001-3098fcf14e08/volumes" Nov 24 17:30:01 crc kubenswrapper[4777]: I1124 17:30:01.264436 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd782d01-363c-4e79-9faa-e5266519cf90" path="/var/lib/kubelet/pods/cd782d01-363c-4e79-9faa-e5266519cf90/volumes" Nov 24 17:30:01 crc kubenswrapper[4777]: I1124 17:30:01.856837 4777 generic.go:334] "Generic (PLEG): container finished" podID="5fbbf0f0-b4c2-4e65-be72-d50116119445" containerID="e3314767e16286e35f07ab67654e0578da920c7b5229fb2243ac88d64d00577c" exitCode=0 Nov 24 17:30:01 crc kubenswrapper[4777]: I1124 17:30:01.856877 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" event={"ID":"5fbbf0f0-b4c2-4e65-be72-d50116119445","Type":"ContainerDied","Data":"e3314767e16286e35f07ab67654e0578da920c7b5229fb2243ac88d64d00577c"} Nov 24 17:30:01 crc kubenswrapper[4777]: I1124 17:30:01.856902 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" event={"ID":"5fbbf0f0-b4c2-4e65-be72-d50116119445","Type":"ContainerStarted","Data":"70f76ab6a0649ce43ca76a3a51ffbfb60148b22691f8c44afed24ec400441684"} Nov 24 17:30:02 crc kubenswrapper[4777]: I1124 17:30:02.028926 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7311-account-create-xpsjd"] Nov 24 17:30:02 crc kubenswrapper[4777]: I1124 17:30:02.036842 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-jdldd"] Nov 24 17:30:02 crc kubenswrapper[4777]: I1124 17:30:02.046365 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-75wbk"] Nov 24 17:30:02 crc kubenswrapper[4777]: I1124 17:30:02.055085 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-75wbk"] Nov 24 17:30:02 crc kubenswrapper[4777]: I1124 17:30:02.064162 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7311-account-create-xpsjd"] Nov 24 17:30:02 crc kubenswrapper[4777]: I1124 17:30:02.077157 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-jdldd"] Nov 24 17:30:02 crc kubenswrapper[4777]: I1124 17:30:02.245637 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:30:02 crc kubenswrapper[4777]: E1124 17:30:02.246425 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.262320 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a11e0baf-fea0-409a-9a16-1a467f83c1bf" path="/var/lib/kubelet/pods/a11e0baf-fea0-409a-9a16-1a467f83c1bf/volumes" Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.263169 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baaece26-e676-4fcd-a0b2-fd4202d7631a" path="/var/lib/kubelet/pods/baaece26-e676-4fcd-a0b2-fd4202d7631a/volumes" Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.263719 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f254c75d-4cc6-43b0-8f88-517f7539d2a5" path="/var/lib/kubelet/pods/f254c75d-4cc6-43b0-8f88-517f7539d2a5/volumes" Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.323128 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.462172 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwf5x\" (UniqueName: \"kubernetes.io/projected/5fbbf0f0-b4c2-4e65-be72-d50116119445-kube-api-access-wwf5x\") pod \"5fbbf0f0-b4c2-4e65-be72-d50116119445\" (UID: \"5fbbf0f0-b4c2-4e65-be72-d50116119445\") " Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.462527 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fbbf0f0-b4c2-4e65-be72-d50116119445-secret-volume\") pod \"5fbbf0f0-b4c2-4e65-be72-d50116119445\" (UID: \"5fbbf0f0-b4c2-4e65-be72-d50116119445\") " Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.462599 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fbbf0f0-b4c2-4e65-be72-d50116119445-config-volume\") pod \"5fbbf0f0-b4c2-4e65-be72-d50116119445\" (UID: \"5fbbf0f0-b4c2-4e65-be72-d50116119445\") " Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.463776 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fbbf0f0-b4c2-4e65-be72-d50116119445-config-volume" (OuterVolumeSpecName: "config-volume") pod "5fbbf0f0-b4c2-4e65-be72-d50116119445" (UID: "5fbbf0f0-b4c2-4e65-be72-d50116119445"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.468353 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fbbf0f0-b4c2-4e65-be72-d50116119445-kube-api-access-wwf5x" (OuterVolumeSpecName: "kube-api-access-wwf5x") pod "5fbbf0f0-b4c2-4e65-be72-d50116119445" (UID: "5fbbf0f0-b4c2-4e65-be72-d50116119445"). InnerVolumeSpecName "kube-api-access-wwf5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.478817 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fbbf0f0-b4c2-4e65-be72-d50116119445-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5fbbf0f0-b4c2-4e65-be72-d50116119445" (UID: "5fbbf0f0-b4c2-4e65-be72-d50116119445"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.565632 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwf5x\" (UniqueName: \"kubernetes.io/projected/5fbbf0f0-b4c2-4e65-be72-d50116119445-kube-api-access-wwf5x\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.565917 4777 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fbbf0f0-b4c2-4e65-be72-d50116119445-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.566044 4777 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fbbf0f0-b4c2-4e65-be72-d50116119445-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.879573 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" event={"ID":"5fbbf0f0-b4c2-4e65-be72-d50116119445","Type":"ContainerDied","Data":"70f76ab6a0649ce43ca76a3a51ffbfb60148b22691f8c44afed24ec400441684"} Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.879937 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70f76ab6a0649ce43ca76a3a51ffbfb60148b22691f8c44afed24ec400441684" Nov 24 17:30:03 crc kubenswrapper[4777]: I1124 17:30:03.879655 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400090-vwx4m" Nov 24 17:30:15 crc kubenswrapper[4777]: I1124 17:30:15.245201 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:30:15 crc kubenswrapper[4777]: E1124 17:30:15.246225 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:30:28 crc kubenswrapper[4777]: I1124 17:30:28.058445 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-5d5f-account-create-vxk8c"] Nov 24 17:30:28 crc kubenswrapper[4777]: I1124 17:30:28.074856 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-c352-account-create-r4cp5"] Nov 24 17:30:28 crc kubenswrapper[4777]: I1124 17:30:28.086126 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-19af-account-create-9fl7d"] Nov 24 17:30:28 crc kubenswrapper[4777]: I1124 17:30:28.102928 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-5d5f-account-create-vxk8c"] Nov 24 17:30:28 crc kubenswrapper[4777]: I1124 17:30:28.110395 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-9666p"] Nov 24 17:30:28 crc kubenswrapper[4777]: I1124 17:30:28.118681 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-q9djb"] Nov 24 17:30:28 crc kubenswrapper[4777]: I1124 17:30:28.126813 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-590a-account-create-kxt2z"] Nov 24 17:30:28 crc kubenswrapper[4777]: I1124 17:30:28.134847 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-19af-account-create-9fl7d"] Nov 24 17:30:28 crc kubenswrapper[4777]: I1124 17:30:28.142536 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-fzsq6"] Nov 24 17:30:28 crc kubenswrapper[4777]: I1124 17:30:28.151229 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-c352-account-create-r4cp5"] Nov 24 17:30:28 crc kubenswrapper[4777]: I1124 17:30:28.159432 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-q9djb"] Nov 24 17:30:28 crc kubenswrapper[4777]: I1124 17:30:28.167701 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-590a-account-create-kxt2z"] Nov 24 17:30:28 crc kubenswrapper[4777]: I1124 17:30:28.177233 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-fzsq6"] Nov 24 17:30:28 crc kubenswrapper[4777]: I1124 17:30:28.185191 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-9666p"] Nov 24 17:30:28 crc kubenswrapper[4777]: I1124 17:30:28.245333 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:30:28 crc kubenswrapper[4777]: E1124 17:30:28.245791 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:30:29 crc kubenswrapper[4777]: I1124 17:30:29.262830 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c41309a-1061-4cf2-bf89-f170ee446905" path="/var/lib/kubelet/pods/7c41309a-1061-4cf2-bf89-f170ee446905/volumes" Nov 24 17:30:29 crc kubenswrapper[4777]: I1124 17:30:29.264033 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81ec0e3a-1906-4013-a5cd-c639933a26a5" path="/var/lib/kubelet/pods/81ec0e3a-1906-4013-a5cd-c639933a26a5/volumes" Nov 24 17:30:29 crc kubenswrapper[4777]: I1124 17:30:29.265152 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d7c6a9d-cea0-4fab-abc8-53e1926587cc" path="/var/lib/kubelet/pods/8d7c6a9d-cea0-4fab-abc8-53e1926587cc/volumes" Nov 24 17:30:29 crc kubenswrapper[4777]: I1124 17:30:29.266223 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9721c23b-c04c-4dc4-94ab-24455982fbae" path="/var/lib/kubelet/pods/9721c23b-c04c-4dc4-94ab-24455982fbae/volumes" Nov 24 17:30:29 crc kubenswrapper[4777]: I1124 17:30:29.269451 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab4c523b-a2f8-4c22-bbf8-c4ce606bd932" path="/var/lib/kubelet/pods/ab4c523b-a2f8-4c22-bbf8-c4ce606bd932/volumes" Nov 24 17:30:29 crc kubenswrapper[4777]: I1124 17:30:29.275144 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6389225-005e-4032-a3bb-a5700e55d08e" path="/var/lib/kubelet/pods/c6389225-005e-4032-a3bb-a5700e55d08e/volumes" Nov 24 17:30:29 crc kubenswrapper[4777]: I1124 17:30:29.277375 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e082ddb3-c5ea-4e65-b002-abefd538a16b" path="/var/lib/kubelet/pods/e082ddb3-c5ea-4e65-b002-abefd538a16b/volumes" Nov 24 17:30:31 crc kubenswrapper[4777]: I1124 17:30:31.028553 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-create-86vqg"] Nov 24 17:30:31 crc kubenswrapper[4777]: I1124 17:30:31.038618 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-create-86vqg"] Nov 24 17:30:31 crc kubenswrapper[4777]: I1124 17:30:31.258584 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf61c287-5b51-4b32-a50e-54b93835632d" path="/var/lib/kubelet/pods/bf61c287-5b51-4b32-a50e-54b93835632d/volumes" Nov 24 17:30:37 crc kubenswrapper[4777]: I1124 17:30:37.043524 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-swrgj"] Nov 24 17:30:37 crc kubenswrapper[4777]: I1124 17:30:37.055771 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-swrgj"] Nov 24 17:30:37 crc kubenswrapper[4777]: I1124 17:30:37.256104 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b60b793-78a5-406c-9ed6-e582dfba568b" path="/var/lib/kubelet/pods/7b60b793-78a5-406c-9ed6-e582dfba568b/volumes" Nov 24 17:30:38 crc kubenswrapper[4777]: I1124 17:30:38.030626 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-stwcx"] Nov 24 17:30:38 crc kubenswrapper[4777]: I1124 17:30:38.040002 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-stwcx"] Nov 24 17:30:39 crc kubenswrapper[4777]: I1124 17:30:39.267220 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc393164-6269-4238-8eab-294157bd3e0a" path="/var/lib/kubelet/pods/cc393164-6269-4238-8eab-294157bd3e0a/volumes" Nov 24 17:30:43 crc kubenswrapper[4777]: I1124 17:30:43.252427 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:30:43 crc kubenswrapper[4777]: E1124 17:30:43.253022 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.191748 4777 scope.go:117] "RemoveContainer" containerID="0fc4f625856d9a7ad250bfd9e7804dd80ee0c0dacae165c7e60b289b70473a20" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.215760 4777 scope.go:117] "RemoveContainer" containerID="90dfbc6bba00483452e8de630955c422c88de76c8ab125921c52006e19554208" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.293788 4777 scope.go:117] "RemoveContainer" containerID="e52e01882ae7e79941c7f6b8a9a374979301f72a5efcf864645fa966abf14cca" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.344824 4777 scope.go:117] "RemoveContainer" containerID="0bab90472c03fb88ec36c73d3b0f7c4207b59fca59be37a1911ceaf8e7e1a137" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.381119 4777 scope.go:117] "RemoveContainer" containerID="d57f5c5e958e3c77e35416bfd8f87db04e97a311ca0c6689b7decfa910bf7618" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.440528 4777 scope.go:117] "RemoveContainer" containerID="cdc32b64f9d9589e5cf0fc23fbaa454e1bced455393d8456e88baab88c5f3f2c" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.465595 4777 scope.go:117] "RemoveContainer" containerID="bee6b663e803ef6ed74f656ae8ce0b476f0c81e8125c23b32c96b7fd47089a38" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.494153 4777 scope.go:117] "RemoveContainer" containerID="1941bfbad606f73fef69beaa7014dccf63aa6793a0cc26746ac46ba0082e8e85" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.535996 4777 scope.go:117] "RemoveContainer" containerID="93cccd55686027e03ab1720528be46272682e0d0b294a31e5f9fe1534463b6c7" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.571281 4777 scope.go:117] "RemoveContainer" containerID="57063a5842a346b840a1e19cfdac14c41890e1df1a569da8e9888af51709f7fa" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.637039 4777 scope.go:117] "RemoveContainer" containerID="14ed858c11c45e2627d4e0b296f218ce37b196de7b5e5e448f4796979f0aea26" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.670821 4777 scope.go:117] "RemoveContainer" containerID="c493219defb7ac0b49f6571834ff1d085dd31a86ddbfc2716ef1d14a29eb16b0" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.694538 4777 scope.go:117] "RemoveContainer" containerID="1f61570c43f2b5598bdc200b72f8b7fa54e80489f1203a96ca56ea88f7b25495" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.718325 4777 scope.go:117] "RemoveContainer" containerID="16914b569eb76eb8eec42c98f9b054e6f10324333886c0bba58e9c4445b0e1f6" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.739759 4777 scope.go:117] "RemoveContainer" containerID="d279e716a60bd53020eb206467a900871b6af1b0724000a7ad1429e6114d7844" Nov 24 17:30:46 crc kubenswrapper[4777]: I1124 17:30:46.773677 4777 scope.go:117] "RemoveContainer" containerID="9d7cc59fd3983e9eeb50f7267b2de2bf999330eb8d1ba808e190de2c57944d28" Nov 24 17:30:54 crc kubenswrapper[4777]: I1124 17:30:54.245426 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:30:54 crc kubenswrapper[4777]: E1124 17:30:54.246178 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:30:59 crc kubenswrapper[4777]: I1124 17:30:59.541647 4777 generic.go:334] "Generic (PLEG): container finished" podID="266eceac-bfd7-4597-b218-62cea462ce5c" containerID="7580d122f9bad9d354197b2b9370d1a92a248cd12f7d0d09e9c1b3afcec029d4" exitCode=0 Nov 24 17:30:59 crc kubenswrapper[4777]: I1124 17:30:59.541729 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" event={"ID":"266eceac-bfd7-4597-b218-62cea462ce5c","Type":"ContainerDied","Data":"7580d122f9bad9d354197b2b9370d1a92a248cd12f7d0d09e9c1b3afcec029d4"} Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.131225 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.196758 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntgrj\" (UniqueName: \"kubernetes.io/projected/266eceac-bfd7-4597-b218-62cea462ce5c-kube-api-access-ntgrj\") pod \"266eceac-bfd7-4597-b218-62cea462ce5c\" (UID: \"266eceac-bfd7-4597-b218-62cea462ce5c\") " Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.196909 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/266eceac-bfd7-4597-b218-62cea462ce5c-ssh-key\") pod \"266eceac-bfd7-4597-b218-62cea462ce5c\" (UID: \"266eceac-bfd7-4597-b218-62cea462ce5c\") " Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.197074 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/266eceac-bfd7-4597-b218-62cea462ce5c-inventory\") pod \"266eceac-bfd7-4597-b218-62cea462ce5c\" (UID: \"266eceac-bfd7-4597-b218-62cea462ce5c\") " Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.204233 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/266eceac-bfd7-4597-b218-62cea462ce5c-kube-api-access-ntgrj" (OuterVolumeSpecName: "kube-api-access-ntgrj") pod "266eceac-bfd7-4597-b218-62cea462ce5c" (UID: "266eceac-bfd7-4597-b218-62cea462ce5c"). InnerVolumeSpecName "kube-api-access-ntgrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.229603 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/266eceac-bfd7-4597-b218-62cea462ce5c-inventory" (OuterVolumeSpecName: "inventory") pod "266eceac-bfd7-4597-b218-62cea462ce5c" (UID: "266eceac-bfd7-4597-b218-62cea462ce5c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.238629 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/266eceac-bfd7-4597-b218-62cea462ce5c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "266eceac-bfd7-4597-b218-62cea462ce5c" (UID: "266eceac-bfd7-4597-b218-62cea462ce5c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.299816 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntgrj\" (UniqueName: \"kubernetes.io/projected/266eceac-bfd7-4597-b218-62cea462ce5c-kube-api-access-ntgrj\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.299850 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/266eceac-bfd7-4597-b218-62cea462ce5c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.299862 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/266eceac-bfd7-4597-b218-62cea462ce5c-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.588158 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" event={"ID":"266eceac-bfd7-4597-b218-62cea462ce5c","Type":"ContainerDied","Data":"aff38be9e6432737eecd36e8cf41e3aafa02764783da9ea0216eaf1cb85dd6a3"} Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.588570 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aff38be9e6432737eecd36e8cf41e3aafa02764783da9ea0216eaf1cb85dd6a3" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.588261 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.669836 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz"] Nov 24 17:31:01 crc kubenswrapper[4777]: E1124 17:31:01.670644 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="266eceac-bfd7-4597-b218-62cea462ce5c" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.670679 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="266eceac-bfd7-4597-b218-62cea462ce5c" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 17:31:01 crc kubenswrapper[4777]: E1124 17:31:01.670696 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fbbf0f0-b4c2-4e65-be72-d50116119445" containerName="collect-profiles" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.670704 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fbbf0f0-b4c2-4e65-be72-d50116119445" containerName="collect-profiles" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.683167 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="266eceac-bfd7-4597-b218-62cea462ce5c" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.683224 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fbbf0f0-b4c2-4e65-be72-d50116119445" containerName="collect-profiles" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.684393 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.688115 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.688780 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.688942 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.689098 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.692516 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz"] Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.711514 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bmrf\" (UniqueName: \"kubernetes.io/projected/45884cb1-7d6c-47a3-b64f-939c1119d70b-kube-api-access-2bmrf\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz\" (UID: \"45884cb1-7d6c-47a3-b64f-939c1119d70b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.711655 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/45884cb1-7d6c-47a3-b64f-939c1119d70b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz\" (UID: \"45884cb1-7d6c-47a3-b64f-939c1119d70b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.711725 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/45884cb1-7d6c-47a3-b64f-939c1119d70b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz\" (UID: \"45884cb1-7d6c-47a3-b64f-939c1119d70b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.813104 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/45884cb1-7d6c-47a3-b64f-939c1119d70b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz\" (UID: \"45884cb1-7d6c-47a3-b64f-939c1119d70b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.813202 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/45884cb1-7d6c-47a3-b64f-939c1119d70b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz\" (UID: \"45884cb1-7d6c-47a3-b64f-939c1119d70b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.813323 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bmrf\" (UniqueName: \"kubernetes.io/projected/45884cb1-7d6c-47a3-b64f-939c1119d70b-kube-api-access-2bmrf\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz\" (UID: \"45884cb1-7d6c-47a3-b64f-939c1119d70b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.818687 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/45884cb1-7d6c-47a3-b64f-939c1119d70b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz\" (UID: \"45884cb1-7d6c-47a3-b64f-939c1119d70b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.820175 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/45884cb1-7d6c-47a3-b64f-939c1119d70b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz\" (UID: \"45884cb1-7d6c-47a3-b64f-939c1119d70b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" Nov 24 17:31:01 crc kubenswrapper[4777]: I1124 17:31:01.830653 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bmrf\" (UniqueName: \"kubernetes.io/projected/45884cb1-7d6c-47a3-b64f-939c1119d70b-kube-api-access-2bmrf\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz\" (UID: \"45884cb1-7d6c-47a3-b64f-939c1119d70b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" Nov 24 17:31:02 crc kubenswrapper[4777]: I1124 17:31:02.024004 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" Nov 24 17:31:02 crc kubenswrapper[4777]: I1124 17:31:02.655747 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz"] Nov 24 17:31:02 crc kubenswrapper[4777]: I1124 17:31:02.668153 4777 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:31:03 crc kubenswrapper[4777]: I1124 17:31:03.605742 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" event={"ID":"45884cb1-7d6c-47a3-b64f-939c1119d70b","Type":"ContainerStarted","Data":"6b472d87cd40d9952691bd275bf6504303646175eab14c75072c8826e3ed6f2e"} Nov 24 17:31:03 crc kubenswrapper[4777]: I1124 17:31:03.606370 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" event={"ID":"45884cb1-7d6c-47a3-b64f-939c1119d70b","Type":"ContainerStarted","Data":"ab1a3796fb065873b58e87d90c7d4d54230f4cfc15fe84c3f57ad3075468216f"} Nov 24 17:31:03 crc kubenswrapper[4777]: I1124 17:31:03.626982 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" podStartSLOduration=2.140993755 podStartE2EDuration="2.626949476s" podCreationTimestamp="2025-11-24 17:31:01 +0000 UTC" firstStartedPulling="2025-11-24 17:31:02.6678013 +0000 UTC m=+1820.826996359" lastFinishedPulling="2025-11-24 17:31:03.153757031 +0000 UTC m=+1821.312952080" observedRunningTime="2025-11-24 17:31:03.620241914 +0000 UTC m=+1821.779436953" watchObservedRunningTime="2025-11-24 17:31:03.626949476 +0000 UTC m=+1821.786144525" Nov 24 17:31:09 crc kubenswrapper[4777]: I1124 17:31:09.245243 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:31:09 crc kubenswrapper[4777]: E1124 17:31:09.246062 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:31:21 crc kubenswrapper[4777]: I1124 17:31:21.246336 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:31:21 crc kubenswrapper[4777]: I1124 17:31:21.844743 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"a52a61b464393ded65c18d60855e7e57f0df7e2cf82fdbb47caf6735df256f42"} Nov 24 17:31:23 crc kubenswrapper[4777]: I1124 17:31:23.104437 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-kl4mk"] Nov 24 17:31:23 crc kubenswrapper[4777]: I1124 17:31:23.113105 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-kl4mk"] Nov 24 17:31:23 crc kubenswrapper[4777]: I1124 17:31:23.257208 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b01f3e72-974a-4649-84a3-d23609d774a9" path="/var/lib/kubelet/pods/b01f3e72-974a-4649-84a3-d23609d774a9/volumes" Nov 24 17:31:26 crc kubenswrapper[4777]: I1124 17:31:26.039766 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-pwl4w"] Nov 24 17:31:26 crc kubenswrapper[4777]: I1124 17:31:26.059433 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-pwl4w"] Nov 24 17:31:26 crc kubenswrapper[4777]: I1124 17:31:26.069350 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-k48pc"] Nov 24 17:31:26 crc kubenswrapper[4777]: I1124 17:31:26.077824 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-k48pc"] Nov 24 17:31:27 crc kubenswrapper[4777]: I1124 17:31:27.268423 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d2fef26-a53d-41dd-8523-b451f2ea8127" path="/var/lib/kubelet/pods/6d2fef26-a53d-41dd-8523-b451f2ea8127/volumes" Nov 24 17:31:27 crc kubenswrapper[4777]: I1124 17:31:27.270378 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d84e561b-fba7-4438-9959-18e1b4bd889c" path="/var/lib/kubelet/pods/d84e561b-fba7-4438-9959-18e1b4bd889c/volumes" Nov 24 17:31:38 crc kubenswrapper[4777]: I1124 17:31:38.036244 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-v7hr7"] Nov 24 17:31:38 crc kubenswrapper[4777]: I1124 17:31:38.050032 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-v7hr7"] Nov 24 17:31:39 crc kubenswrapper[4777]: I1124 17:31:39.261114 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4151db5-e476-4ffa-a038-369874b439cc" path="/var/lib/kubelet/pods/d4151db5-e476-4ffa-a038-369874b439cc/volumes" Nov 24 17:31:47 crc kubenswrapper[4777]: I1124 17:31:47.093778 4777 scope.go:117] "RemoveContainer" containerID="976ea245350feaf855dc1afa93656b1d0e6f42e14b8b5449334c3e0ea9661d25" Nov 24 17:31:47 crc kubenswrapper[4777]: I1124 17:31:47.128638 4777 scope.go:117] "RemoveContainer" containerID="ea7e7c17d4adfe7017b3c1ba21aa4f3026ea25523f283fc55a8ac4b0acecf7c6" Nov 24 17:31:47 crc kubenswrapper[4777]: I1124 17:31:47.183014 4777 scope.go:117] "RemoveContainer" containerID="935312b5eed46e207be26dfa6e5cc349dffef7c3e3093aa5bf5591c49fdaa41b" Nov 24 17:31:47 crc kubenswrapper[4777]: I1124 17:31:47.249232 4777 scope.go:117] "RemoveContainer" containerID="dc3a5cc17f9fd14cb69378682b5c09769c0067f44ee7ca44e40a7b0caa65a536" Nov 24 17:31:56 crc kubenswrapper[4777]: I1124 17:31:56.038225 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-6wwt4"] Nov 24 17:31:56 crc kubenswrapper[4777]: I1124 17:31:56.048928 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-6wwt4"] Nov 24 17:31:57 crc kubenswrapper[4777]: I1124 17:31:57.261949 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba" path="/var/lib/kubelet/pods/95ccd611-1bbb-4a6d-b31e-3fec5e2fd2ba/volumes" Nov 24 17:32:11 crc kubenswrapper[4777]: I1124 17:32:11.038087 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-7d7z8"] Nov 24 17:32:11 crc kubenswrapper[4777]: I1124 17:32:11.050722 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-7d7z8"] Nov 24 17:32:11 crc kubenswrapper[4777]: I1124 17:32:11.264234 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a121eb7-f7d9-4d42-9dba-38225f102982" path="/var/lib/kubelet/pods/3a121eb7-f7d9-4d42-9dba-38225f102982/volumes" Nov 24 17:32:16 crc kubenswrapper[4777]: I1124 17:32:16.386417 4777 generic.go:334] "Generic (PLEG): container finished" podID="45884cb1-7d6c-47a3-b64f-939c1119d70b" containerID="6b472d87cd40d9952691bd275bf6504303646175eab14c75072c8826e3ed6f2e" exitCode=0 Nov 24 17:32:16 crc kubenswrapper[4777]: I1124 17:32:16.387204 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" event={"ID":"45884cb1-7d6c-47a3-b64f-939c1119d70b","Type":"ContainerDied","Data":"6b472d87cd40d9952691bd275bf6504303646175eab14c75072c8826e3ed6f2e"} Nov 24 17:32:17 crc kubenswrapper[4777]: I1124 17:32:17.908172 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.023358 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bmrf\" (UniqueName: \"kubernetes.io/projected/45884cb1-7d6c-47a3-b64f-939c1119d70b-kube-api-access-2bmrf\") pod \"45884cb1-7d6c-47a3-b64f-939c1119d70b\" (UID: \"45884cb1-7d6c-47a3-b64f-939c1119d70b\") " Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.023599 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/45884cb1-7d6c-47a3-b64f-939c1119d70b-ssh-key\") pod \"45884cb1-7d6c-47a3-b64f-939c1119d70b\" (UID: \"45884cb1-7d6c-47a3-b64f-939c1119d70b\") " Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.023699 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/45884cb1-7d6c-47a3-b64f-939c1119d70b-inventory\") pod \"45884cb1-7d6c-47a3-b64f-939c1119d70b\" (UID: \"45884cb1-7d6c-47a3-b64f-939c1119d70b\") " Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.033063 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45884cb1-7d6c-47a3-b64f-939c1119d70b-kube-api-access-2bmrf" (OuterVolumeSpecName: "kube-api-access-2bmrf") pod "45884cb1-7d6c-47a3-b64f-939c1119d70b" (UID: "45884cb1-7d6c-47a3-b64f-939c1119d70b"). InnerVolumeSpecName "kube-api-access-2bmrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.055556 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45884cb1-7d6c-47a3-b64f-939c1119d70b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "45884cb1-7d6c-47a3-b64f-939c1119d70b" (UID: "45884cb1-7d6c-47a3-b64f-939c1119d70b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.061037 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45884cb1-7d6c-47a3-b64f-939c1119d70b-inventory" (OuterVolumeSpecName: "inventory") pod "45884cb1-7d6c-47a3-b64f-939c1119d70b" (UID: "45884cb1-7d6c-47a3-b64f-939c1119d70b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.127019 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/45884cb1-7d6c-47a3-b64f-939c1119d70b-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.127149 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/45884cb1-7d6c-47a3-b64f-939c1119d70b-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.127258 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bmrf\" (UniqueName: \"kubernetes.io/projected/45884cb1-7d6c-47a3-b64f-939c1119d70b-kube-api-access-2bmrf\") on node \"crc\" DevicePath \"\"" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.408275 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" event={"ID":"45884cb1-7d6c-47a3-b64f-939c1119d70b","Type":"ContainerDied","Data":"ab1a3796fb065873b58e87d90c7d4d54230f4cfc15fe84c3f57ad3075468216f"} Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.408321 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab1a3796fb065873b58e87d90c7d4d54230f4cfc15fe84c3f57ad3075468216f" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.408338 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.522601 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j"] Nov 24 17:32:18 crc kubenswrapper[4777]: E1124 17:32:18.523329 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45884cb1-7d6c-47a3-b64f-939c1119d70b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.523358 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="45884cb1-7d6c-47a3-b64f-939c1119d70b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.523801 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="45884cb1-7d6c-47a3-b64f-939c1119d70b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.524943 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.527714 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.527739 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.527716 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.527893 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.539564 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j"] Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.638540 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/541aa87c-3013-4641-950b-a23f825bcb96-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j\" (UID: \"541aa87c-3013-4641-950b-a23f825bcb96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.638599 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chlds\" (UniqueName: \"kubernetes.io/projected/541aa87c-3013-4641-950b-a23f825bcb96-kube-api-access-chlds\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j\" (UID: \"541aa87c-3013-4641-950b-a23f825bcb96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.638622 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/541aa87c-3013-4641-950b-a23f825bcb96-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j\" (UID: \"541aa87c-3013-4641-950b-a23f825bcb96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.741385 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chlds\" (UniqueName: \"kubernetes.io/projected/541aa87c-3013-4641-950b-a23f825bcb96-kube-api-access-chlds\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j\" (UID: \"541aa87c-3013-4641-950b-a23f825bcb96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.741450 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/541aa87c-3013-4641-950b-a23f825bcb96-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j\" (UID: \"541aa87c-3013-4641-950b-a23f825bcb96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.741702 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/541aa87c-3013-4641-950b-a23f825bcb96-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j\" (UID: \"541aa87c-3013-4641-950b-a23f825bcb96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.746704 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/541aa87c-3013-4641-950b-a23f825bcb96-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j\" (UID: \"541aa87c-3013-4641-950b-a23f825bcb96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.748112 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/541aa87c-3013-4641-950b-a23f825bcb96-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j\" (UID: \"541aa87c-3013-4641-950b-a23f825bcb96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.760558 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chlds\" (UniqueName: \"kubernetes.io/projected/541aa87c-3013-4641-950b-a23f825bcb96-kube-api-access-chlds\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j\" (UID: \"541aa87c-3013-4641-950b-a23f825bcb96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" Nov 24 17:32:18 crc kubenswrapper[4777]: I1124 17:32:18.844795 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" Nov 24 17:32:19 crc kubenswrapper[4777]: I1124 17:32:19.524152 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j"] Nov 24 17:32:20 crc kubenswrapper[4777]: I1124 17:32:20.426484 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" event={"ID":"541aa87c-3013-4641-950b-a23f825bcb96","Type":"ContainerStarted","Data":"b64d7dab5907a0450008ac0a9bf2dc903e59c5c31594538d805d0f04c37b9aab"} Nov 24 17:32:20 crc kubenswrapper[4777]: I1124 17:32:20.427276 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" event={"ID":"541aa87c-3013-4641-950b-a23f825bcb96","Type":"ContainerStarted","Data":"5fb3888b557338c17837a71ac4f6494ec6ef6262190f434ee8e65d61381cb491"} Nov 24 17:32:20 crc kubenswrapper[4777]: I1124 17:32:20.460530 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" podStartSLOduration=2.040044157 podStartE2EDuration="2.460511213s" podCreationTimestamp="2025-11-24 17:32:18 +0000 UTC" firstStartedPulling="2025-11-24 17:32:19.533679722 +0000 UTC m=+1897.692874771" lastFinishedPulling="2025-11-24 17:32:19.954146778 +0000 UTC m=+1898.113341827" observedRunningTime="2025-11-24 17:32:20.459310609 +0000 UTC m=+1898.618505648" watchObservedRunningTime="2025-11-24 17:32:20.460511213 +0000 UTC m=+1898.619706262" Nov 24 17:32:21 crc kubenswrapper[4777]: I1124 17:32:21.037247 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-xshsz"] Nov 24 17:32:21 crc kubenswrapper[4777]: I1124 17:32:21.046235 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-18da-account-create-swzmb"] Nov 24 17:32:21 crc kubenswrapper[4777]: I1124 17:32:21.055226 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-xshsz"] Nov 24 17:32:21 crc kubenswrapper[4777]: I1124 17:32:21.065620 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-18da-account-create-swzmb"] Nov 24 17:32:21 crc kubenswrapper[4777]: I1124 17:32:21.075541 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-tpnnn"] Nov 24 17:32:21 crc kubenswrapper[4777]: I1124 17:32:21.083452 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-tpnnn"] Nov 24 17:32:21 crc kubenswrapper[4777]: I1124 17:32:21.092861 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-287c-account-create-gnrdg"] Nov 24 17:32:21 crc kubenswrapper[4777]: I1124 17:32:21.100709 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-a9bb-account-create-5d494"] Nov 24 17:32:21 crc kubenswrapper[4777]: I1124 17:32:21.108574 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-a9bb-account-create-5d494"] Nov 24 17:32:21 crc kubenswrapper[4777]: I1124 17:32:21.116324 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-287c-account-create-gnrdg"] Nov 24 17:32:21 crc kubenswrapper[4777]: I1124 17:32:21.256916 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="051bfdbe-233c-47db-a89f-ff3c89cb342a" path="/var/lib/kubelet/pods/051bfdbe-233c-47db-a89f-ff3c89cb342a/volumes" Nov 24 17:32:21 crc kubenswrapper[4777]: I1124 17:32:21.258018 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56" path="/var/lib/kubelet/pods/2b4460b1-c7fa-4b18-b6f9-7a0c092b8f56/volumes" Nov 24 17:32:21 crc kubenswrapper[4777]: I1124 17:32:21.258574 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="762d2534-27f5-4323-882d-24f794b2034b" path="/var/lib/kubelet/pods/762d2534-27f5-4323-882d-24f794b2034b/volumes" Nov 24 17:32:21 crc kubenswrapper[4777]: I1124 17:32:21.259116 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cbe8193-bfdb-44bb-94af-21372139923c" path="/var/lib/kubelet/pods/7cbe8193-bfdb-44bb-94af-21372139923c/volumes" Nov 24 17:32:21 crc kubenswrapper[4777]: I1124 17:32:21.260137 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d273d56f-f421-4a2b-8dd5-faaf7603ee79" path="/var/lib/kubelet/pods/d273d56f-f421-4a2b-8dd5-faaf7603ee79/volumes" Nov 24 17:32:25 crc kubenswrapper[4777]: I1124 17:32:25.490894 4777 generic.go:334] "Generic (PLEG): container finished" podID="541aa87c-3013-4641-950b-a23f825bcb96" containerID="b64d7dab5907a0450008ac0a9bf2dc903e59c5c31594538d805d0f04c37b9aab" exitCode=0 Nov 24 17:32:25 crc kubenswrapper[4777]: I1124 17:32:25.491049 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" event={"ID":"541aa87c-3013-4641-950b-a23f825bcb96","Type":"ContainerDied","Data":"b64d7dab5907a0450008ac0a9bf2dc903e59c5c31594538d805d0f04c37b9aab"} Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.011530 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.208144 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chlds\" (UniqueName: \"kubernetes.io/projected/541aa87c-3013-4641-950b-a23f825bcb96-kube-api-access-chlds\") pod \"541aa87c-3013-4641-950b-a23f825bcb96\" (UID: \"541aa87c-3013-4641-950b-a23f825bcb96\") " Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.208648 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/541aa87c-3013-4641-950b-a23f825bcb96-ssh-key\") pod \"541aa87c-3013-4641-950b-a23f825bcb96\" (UID: \"541aa87c-3013-4641-950b-a23f825bcb96\") " Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.208860 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/541aa87c-3013-4641-950b-a23f825bcb96-inventory\") pod \"541aa87c-3013-4641-950b-a23f825bcb96\" (UID: \"541aa87c-3013-4641-950b-a23f825bcb96\") " Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.220878 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/541aa87c-3013-4641-950b-a23f825bcb96-kube-api-access-chlds" (OuterVolumeSpecName: "kube-api-access-chlds") pod "541aa87c-3013-4641-950b-a23f825bcb96" (UID: "541aa87c-3013-4641-950b-a23f825bcb96"). InnerVolumeSpecName "kube-api-access-chlds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.245586 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/541aa87c-3013-4641-950b-a23f825bcb96-inventory" (OuterVolumeSpecName: "inventory") pod "541aa87c-3013-4641-950b-a23f825bcb96" (UID: "541aa87c-3013-4641-950b-a23f825bcb96"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.257080 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/541aa87c-3013-4641-950b-a23f825bcb96-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "541aa87c-3013-4641-950b-a23f825bcb96" (UID: "541aa87c-3013-4641-950b-a23f825bcb96"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.311532 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/541aa87c-3013-4641-950b-a23f825bcb96-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.311569 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/541aa87c-3013-4641-950b-a23f825bcb96-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.311586 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chlds\" (UniqueName: \"kubernetes.io/projected/541aa87c-3013-4641-950b-a23f825bcb96-kube-api-access-chlds\") on node \"crc\" DevicePath \"\"" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.523500 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" event={"ID":"541aa87c-3013-4641-950b-a23f825bcb96","Type":"ContainerDied","Data":"5fb3888b557338c17837a71ac4f6494ec6ef6262190f434ee8e65d61381cb491"} Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.523582 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fb3888b557338c17837a71ac4f6494ec6ef6262190f434ee8e65d61381cb491" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.523599 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.636958 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6"] Nov 24 17:32:27 crc kubenswrapper[4777]: E1124 17:32:27.637383 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="541aa87c-3013-4641-950b-a23f825bcb96" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.637400 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="541aa87c-3013-4641-950b-a23f825bcb96" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.637583 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="541aa87c-3013-4641-950b-a23f825bcb96" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.638299 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.641426 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.642201 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.643013 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.643567 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.662166 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6"] Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.720076 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/346142b6-2a63-4610-b1d0-bfebac8c7c61-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rvlp6\" (UID: \"346142b6-2a63-4610-b1d0-bfebac8c7c61\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.720182 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/346142b6-2a63-4610-b1d0-bfebac8c7c61-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rvlp6\" (UID: \"346142b6-2a63-4610-b1d0-bfebac8c7c61\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.720213 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpzcp\" (UniqueName: \"kubernetes.io/projected/346142b6-2a63-4610-b1d0-bfebac8c7c61-kube-api-access-rpzcp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rvlp6\" (UID: \"346142b6-2a63-4610-b1d0-bfebac8c7c61\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.821986 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/346142b6-2a63-4610-b1d0-bfebac8c7c61-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rvlp6\" (UID: \"346142b6-2a63-4610-b1d0-bfebac8c7c61\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.822047 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpzcp\" (UniqueName: \"kubernetes.io/projected/346142b6-2a63-4610-b1d0-bfebac8c7c61-kube-api-access-rpzcp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rvlp6\" (UID: \"346142b6-2a63-4610-b1d0-bfebac8c7c61\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.822194 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/346142b6-2a63-4610-b1d0-bfebac8c7c61-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rvlp6\" (UID: \"346142b6-2a63-4610-b1d0-bfebac8c7c61\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.834742 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/346142b6-2a63-4610-b1d0-bfebac8c7c61-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rvlp6\" (UID: \"346142b6-2a63-4610-b1d0-bfebac8c7c61\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.839757 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpzcp\" (UniqueName: \"kubernetes.io/projected/346142b6-2a63-4610-b1d0-bfebac8c7c61-kube-api-access-rpzcp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rvlp6\" (UID: \"346142b6-2a63-4610-b1d0-bfebac8c7c61\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.878527 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/346142b6-2a63-4610-b1d0-bfebac8c7c61-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rvlp6\" (UID: \"346142b6-2a63-4610-b1d0-bfebac8c7c61\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" Nov 24 17:32:27 crc kubenswrapper[4777]: I1124 17:32:27.960562 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" Nov 24 17:32:28 crc kubenswrapper[4777]: I1124 17:32:28.037800 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wv2r4"] Nov 24 17:32:28 crc kubenswrapper[4777]: I1124 17:32:28.040590 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:28 crc kubenswrapper[4777]: I1124 17:32:28.050899 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wv2r4"] Nov 24 17:32:28 crc kubenswrapper[4777]: I1124 17:32:28.127815 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-catalog-content\") pod \"certified-operators-wv2r4\" (UID: \"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a\") " pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:28 crc kubenswrapper[4777]: I1124 17:32:28.128133 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-utilities\") pod \"certified-operators-wv2r4\" (UID: \"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a\") " pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:28 crc kubenswrapper[4777]: I1124 17:32:28.128442 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9thbm\" (UniqueName: \"kubernetes.io/projected/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-kube-api-access-9thbm\") pod \"certified-operators-wv2r4\" (UID: \"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a\") " pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:28 crc kubenswrapper[4777]: I1124 17:32:28.230836 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9thbm\" (UniqueName: \"kubernetes.io/projected/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-kube-api-access-9thbm\") pod \"certified-operators-wv2r4\" (UID: \"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a\") " pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:28 crc kubenswrapper[4777]: I1124 17:32:28.230906 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-catalog-content\") pod \"certified-operators-wv2r4\" (UID: \"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a\") " pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:28 crc kubenswrapper[4777]: I1124 17:32:28.231056 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-utilities\") pod \"certified-operators-wv2r4\" (UID: \"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a\") " pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:28 crc kubenswrapper[4777]: I1124 17:32:28.231656 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-utilities\") pod \"certified-operators-wv2r4\" (UID: \"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a\") " pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:28 crc kubenswrapper[4777]: I1124 17:32:28.231750 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-catalog-content\") pod \"certified-operators-wv2r4\" (UID: \"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a\") " pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:28 crc kubenswrapper[4777]: I1124 17:32:28.380950 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9thbm\" (UniqueName: \"kubernetes.io/projected/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-kube-api-access-9thbm\") pod \"certified-operators-wv2r4\" (UID: \"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a\") " pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:28 crc kubenswrapper[4777]: I1124 17:32:28.403380 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:28 crc kubenswrapper[4777]: I1124 17:32:28.968452 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6"] Nov 24 17:32:29 crc kubenswrapper[4777]: I1124 17:32:29.039446 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wv2r4"] Nov 24 17:32:29 crc kubenswrapper[4777]: W1124 17:32:29.042285 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a4825d4_adb7_4f1b_b51b_fe860ad98e3a.slice/crio-ebe205b31e7a12064cbf529ca08b338a0d9ffad6eb54fda931fe70220b1adee5 WatchSource:0}: Error finding container ebe205b31e7a12064cbf529ca08b338a0d9ffad6eb54fda931fe70220b1adee5: Status 404 returned error can't find the container with id ebe205b31e7a12064cbf529ca08b338a0d9ffad6eb54fda931fe70220b1adee5 Nov 24 17:32:29 crc kubenswrapper[4777]: I1124 17:32:29.556046 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" event={"ID":"346142b6-2a63-4610-b1d0-bfebac8c7c61","Type":"ContainerStarted","Data":"a128a35d66debf0eecc8eeca9a7eb8d8203fd31f0f40b8e25bb1ada099d7ff8b"} Nov 24 17:32:29 crc kubenswrapper[4777]: I1124 17:32:29.558478 4777 generic.go:334] "Generic (PLEG): container finished" podID="0a4825d4-adb7-4f1b-b51b-fe860ad98e3a" containerID="fce8ed0b09a61306fb279bdc693486c21f677259042fbf5f3cf7225fadd9856d" exitCode=0 Nov 24 17:32:29 crc kubenswrapper[4777]: I1124 17:32:29.558530 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wv2r4" event={"ID":"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a","Type":"ContainerDied","Data":"fce8ed0b09a61306fb279bdc693486c21f677259042fbf5f3cf7225fadd9856d"} Nov 24 17:32:29 crc kubenswrapper[4777]: I1124 17:32:29.558578 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wv2r4" event={"ID":"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a","Type":"ContainerStarted","Data":"ebe205b31e7a12064cbf529ca08b338a0d9ffad6eb54fda931fe70220b1adee5"} Nov 24 17:32:30 crc kubenswrapper[4777]: I1124 17:32:30.569934 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wv2r4" event={"ID":"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a","Type":"ContainerStarted","Data":"d8608ca0922161cc0366ade1dd61ec571687b1f70ab75b1d5c498c9dff373d0d"} Nov 24 17:32:30 crc kubenswrapper[4777]: I1124 17:32:30.573244 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" event={"ID":"346142b6-2a63-4610-b1d0-bfebac8c7c61","Type":"ContainerStarted","Data":"789677121af355ba47fd439a50cf541572b210b5e3e01b8dc2016426b58faef4"} Nov 24 17:32:30 crc kubenswrapper[4777]: I1124 17:32:30.610527 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" podStartSLOduration=3.134668188 podStartE2EDuration="3.61051122s" podCreationTimestamp="2025-11-24 17:32:27 +0000 UTC" firstStartedPulling="2025-11-24 17:32:28.97874577 +0000 UTC m=+1907.137940819" lastFinishedPulling="2025-11-24 17:32:29.454588802 +0000 UTC m=+1907.613783851" observedRunningTime="2025-11-24 17:32:30.60459783 +0000 UTC m=+1908.763792879" watchObservedRunningTime="2025-11-24 17:32:30.61051122 +0000 UTC m=+1908.769706269" Nov 24 17:32:32 crc kubenswrapper[4777]: I1124 17:32:32.594371 4777 generic.go:334] "Generic (PLEG): container finished" podID="0a4825d4-adb7-4f1b-b51b-fe860ad98e3a" containerID="d8608ca0922161cc0366ade1dd61ec571687b1f70ab75b1d5c498c9dff373d0d" exitCode=0 Nov 24 17:32:32 crc kubenswrapper[4777]: I1124 17:32:32.594702 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wv2r4" event={"ID":"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a","Type":"ContainerDied","Data":"d8608ca0922161cc0366ade1dd61ec571687b1f70ab75b1d5c498c9dff373d0d"} Nov 24 17:32:33 crc kubenswrapper[4777]: I1124 17:32:33.608177 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wv2r4" event={"ID":"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a","Type":"ContainerStarted","Data":"4ab6f17708ca957aec5324f26d452b218095ab86f909bc8baf9d61ff135e236a"} Nov 24 17:32:33 crc kubenswrapper[4777]: I1124 17:32:33.633712 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wv2r4" podStartSLOduration=2.155291119 podStartE2EDuration="5.633685599s" podCreationTimestamp="2025-11-24 17:32:28 +0000 UTC" firstStartedPulling="2025-11-24 17:32:29.560225836 +0000 UTC m=+1907.719420885" lastFinishedPulling="2025-11-24 17:32:33.038620316 +0000 UTC m=+1911.197815365" observedRunningTime="2025-11-24 17:32:33.624949399 +0000 UTC m=+1911.784144448" watchObservedRunningTime="2025-11-24 17:32:33.633685599 +0000 UTC m=+1911.792880648" Nov 24 17:32:38 crc kubenswrapper[4777]: I1124 17:32:38.403712 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:38 crc kubenswrapper[4777]: I1124 17:32:38.404812 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:38 crc kubenswrapper[4777]: I1124 17:32:38.459197 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:38 crc kubenswrapper[4777]: I1124 17:32:38.717830 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:38 crc kubenswrapper[4777]: I1124 17:32:38.774037 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wv2r4"] Nov 24 17:32:40 crc kubenswrapper[4777]: I1124 17:32:40.676699 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wv2r4" podUID="0a4825d4-adb7-4f1b-b51b-fe860ad98e3a" containerName="registry-server" containerID="cri-o://4ab6f17708ca957aec5324f26d452b218095ab86f909bc8baf9d61ff135e236a" gracePeriod=2 Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.254305 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.366032 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9thbm\" (UniqueName: \"kubernetes.io/projected/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-kube-api-access-9thbm\") pod \"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a\" (UID: \"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a\") " Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.366185 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-utilities\") pod \"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a\" (UID: \"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a\") " Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.366258 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-catalog-content\") pod \"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a\" (UID: \"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a\") " Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.367251 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-utilities" (OuterVolumeSpecName: "utilities") pod "0a4825d4-adb7-4f1b-b51b-fe860ad98e3a" (UID: "0a4825d4-adb7-4f1b-b51b-fe860ad98e3a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.367488 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.372104 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-kube-api-access-9thbm" (OuterVolumeSpecName: "kube-api-access-9thbm") pod "0a4825d4-adb7-4f1b-b51b-fe860ad98e3a" (UID: "0a4825d4-adb7-4f1b-b51b-fe860ad98e3a"). InnerVolumeSpecName "kube-api-access-9thbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.417311 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a4825d4-adb7-4f1b-b51b-fe860ad98e3a" (UID: "0a4825d4-adb7-4f1b-b51b-fe860ad98e3a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.469171 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9thbm\" (UniqueName: \"kubernetes.io/projected/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-kube-api-access-9thbm\") on node \"crc\" DevicePath \"\"" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.469202 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.691029 4777 generic.go:334] "Generic (PLEG): container finished" podID="0a4825d4-adb7-4f1b-b51b-fe860ad98e3a" containerID="4ab6f17708ca957aec5324f26d452b218095ab86f909bc8baf9d61ff135e236a" exitCode=0 Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.691085 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wv2r4" event={"ID":"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a","Type":"ContainerDied","Data":"4ab6f17708ca957aec5324f26d452b218095ab86f909bc8baf9d61ff135e236a"} Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.691120 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wv2r4" event={"ID":"0a4825d4-adb7-4f1b-b51b-fe860ad98e3a","Type":"ContainerDied","Data":"ebe205b31e7a12064cbf529ca08b338a0d9ffad6eb54fda931fe70220b1adee5"} Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.691145 4777 scope.go:117] "RemoveContainer" containerID="4ab6f17708ca957aec5324f26d452b218095ab86f909bc8baf9d61ff135e236a" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.691335 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wv2r4" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.734246 4777 scope.go:117] "RemoveContainer" containerID="d8608ca0922161cc0366ade1dd61ec571687b1f70ab75b1d5c498c9dff373d0d" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.738167 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wv2r4"] Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.746287 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wv2r4"] Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.764679 4777 scope.go:117] "RemoveContainer" containerID="fce8ed0b09a61306fb279bdc693486c21f677259042fbf5f3cf7225fadd9856d" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.820703 4777 scope.go:117] "RemoveContainer" containerID="4ab6f17708ca957aec5324f26d452b218095ab86f909bc8baf9d61ff135e236a" Nov 24 17:32:41 crc kubenswrapper[4777]: E1124 17:32:41.821231 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ab6f17708ca957aec5324f26d452b218095ab86f909bc8baf9d61ff135e236a\": container with ID starting with 4ab6f17708ca957aec5324f26d452b218095ab86f909bc8baf9d61ff135e236a not found: ID does not exist" containerID="4ab6f17708ca957aec5324f26d452b218095ab86f909bc8baf9d61ff135e236a" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.821277 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ab6f17708ca957aec5324f26d452b218095ab86f909bc8baf9d61ff135e236a"} err="failed to get container status \"4ab6f17708ca957aec5324f26d452b218095ab86f909bc8baf9d61ff135e236a\": rpc error: code = NotFound desc = could not find container \"4ab6f17708ca957aec5324f26d452b218095ab86f909bc8baf9d61ff135e236a\": container with ID starting with 4ab6f17708ca957aec5324f26d452b218095ab86f909bc8baf9d61ff135e236a not found: ID does not exist" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.821305 4777 scope.go:117] "RemoveContainer" containerID="d8608ca0922161cc0366ade1dd61ec571687b1f70ab75b1d5c498c9dff373d0d" Nov 24 17:32:41 crc kubenswrapper[4777]: E1124 17:32:41.821798 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8608ca0922161cc0366ade1dd61ec571687b1f70ab75b1d5c498c9dff373d0d\": container with ID starting with d8608ca0922161cc0366ade1dd61ec571687b1f70ab75b1d5c498c9dff373d0d not found: ID does not exist" containerID="d8608ca0922161cc0366ade1dd61ec571687b1f70ab75b1d5c498c9dff373d0d" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.821849 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8608ca0922161cc0366ade1dd61ec571687b1f70ab75b1d5c498c9dff373d0d"} err="failed to get container status \"d8608ca0922161cc0366ade1dd61ec571687b1f70ab75b1d5c498c9dff373d0d\": rpc error: code = NotFound desc = could not find container \"d8608ca0922161cc0366ade1dd61ec571687b1f70ab75b1d5c498c9dff373d0d\": container with ID starting with d8608ca0922161cc0366ade1dd61ec571687b1f70ab75b1d5c498c9dff373d0d not found: ID does not exist" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.821877 4777 scope.go:117] "RemoveContainer" containerID="fce8ed0b09a61306fb279bdc693486c21f677259042fbf5f3cf7225fadd9856d" Nov 24 17:32:41 crc kubenswrapper[4777]: E1124 17:32:41.822245 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fce8ed0b09a61306fb279bdc693486c21f677259042fbf5f3cf7225fadd9856d\": container with ID starting with fce8ed0b09a61306fb279bdc693486c21f677259042fbf5f3cf7225fadd9856d not found: ID does not exist" containerID="fce8ed0b09a61306fb279bdc693486c21f677259042fbf5f3cf7225fadd9856d" Nov 24 17:32:41 crc kubenswrapper[4777]: I1124 17:32:41.822277 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fce8ed0b09a61306fb279bdc693486c21f677259042fbf5f3cf7225fadd9856d"} err="failed to get container status \"fce8ed0b09a61306fb279bdc693486c21f677259042fbf5f3cf7225fadd9856d\": rpc error: code = NotFound desc = could not find container \"fce8ed0b09a61306fb279bdc693486c21f677259042fbf5f3cf7225fadd9856d\": container with ID starting with fce8ed0b09a61306fb279bdc693486c21f677259042fbf5f3cf7225fadd9856d not found: ID does not exist" Nov 24 17:32:43 crc kubenswrapper[4777]: I1124 17:32:43.262752 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a4825d4-adb7-4f1b-b51b-fe860ad98e3a" path="/var/lib/kubelet/pods/0a4825d4-adb7-4f1b-b51b-fe860ad98e3a/volumes" Nov 24 17:32:47 crc kubenswrapper[4777]: I1124 17:32:47.386217 4777 scope.go:117] "RemoveContainer" containerID="a81bea5b353663e05407a4b03afb1b05194ce88382029dba7dc2d427b5606e5b" Nov 24 17:32:47 crc kubenswrapper[4777]: I1124 17:32:47.427331 4777 scope.go:117] "RemoveContainer" containerID="24e8514d456e8b25c8cf1b4c6dab4f462f6a8a89c1f1f6a5d588c27cef0bbfd4" Nov 24 17:32:47 crc kubenswrapper[4777]: I1124 17:32:47.483706 4777 scope.go:117] "RemoveContainer" containerID="4f009c3fbaa25f1413fc6b2a70e9e1eb560a5cfe85b5bec54104bdd1f4012aca" Nov 24 17:32:47 crc kubenswrapper[4777]: I1124 17:32:47.532915 4777 scope.go:117] "RemoveContainer" containerID="5baef7c26317fe983dd13937a633921b0e460e0cba388302662a30cfd101fdfd" Nov 24 17:32:47 crc kubenswrapper[4777]: I1124 17:32:47.588378 4777 scope.go:117] "RemoveContainer" containerID="955033450ae4da4d2e8d7c8e6e68d31e7b2d1cf85c8de5723855f4d63c00ce7b" Nov 24 17:32:47 crc kubenswrapper[4777]: I1124 17:32:47.636728 4777 scope.go:117] "RemoveContainer" containerID="e8ecd53f8e21a9ac85fadb81b7f325d9fab0bc5ac6db8d3bb92bf494b87208f4" Nov 24 17:32:47 crc kubenswrapper[4777]: I1124 17:32:47.754181 4777 scope.go:117] "RemoveContainer" containerID="0841bed172a9a98f66e2c8b89e20550ea7d92eca0caf45afeb2ab6f81479ab19" Nov 24 17:32:57 crc kubenswrapper[4777]: I1124 17:32:57.045942 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5rcr4"] Nov 24 17:32:57 crc kubenswrapper[4777]: I1124 17:32:57.055576 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5rcr4"] Nov 24 17:32:57 crc kubenswrapper[4777]: I1124 17:32:57.261007 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abf0902b-bd70-49df-966b-4cabf47f75df" path="/var/lib/kubelet/pods/abf0902b-bd70-49df-966b-4cabf47f75df/volumes" Nov 24 17:33:05 crc kubenswrapper[4777]: I1124 17:33:05.987315 4777 generic.go:334] "Generic (PLEG): container finished" podID="346142b6-2a63-4610-b1d0-bfebac8c7c61" containerID="789677121af355ba47fd439a50cf541572b210b5e3e01b8dc2016426b58faef4" exitCode=0 Nov 24 17:33:05 crc kubenswrapper[4777]: I1124 17:33:05.987557 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" event={"ID":"346142b6-2a63-4610-b1d0-bfebac8c7c61","Type":"ContainerDied","Data":"789677121af355ba47fd439a50cf541572b210b5e3e01b8dc2016426b58faef4"} Nov 24 17:33:07 crc kubenswrapper[4777]: I1124 17:33:07.583363 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" Nov 24 17:33:07 crc kubenswrapper[4777]: I1124 17:33:07.660747 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/346142b6-2a63-4610-b1d0-bfebac8c7c61-inventory\") pod \"346142b6-2a63-4610-b1d0-bfebac8c7c61\" (UID: \"346142b6-2a63-4610-b1d0-bfebac8c7c61\") " Nov 24 17:33:07 crc kubenswrapper[4777]: I1124 17:33:07.660801 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/346142b6-2a63-4610-b1d0-bfebac8c7c61-ssh-key\") pod \"346142b6-2a63-4610-b1d0-bfebac8c7c61\" (UID: \"346142b6-2a63-4610-b1d0-bfebac8c7c61\") " Nov 24 17:33:07 crc kubenswrapper[4777]: I1124 17:33:07.660900 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpzcp\" (UniqueName: \"kubernetes.io/projected/346142b6-2a63-4610-b1d0-bfebac8c7c61-kube-api-access-rpzcp\") pod \"346142b6-2a63-4610-b1d0-bfebac8c7c61\" (UID: \"346142b6-2a63-4610-b1d0-bfebac8c7c61\") " Nov 24 17:33:07 crc kubenswrapper[4777]: I1124 17:33:07.668494 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/346142b6-2a63-4610-b1d0-bfebac8c7c61-kube-api-access-rpzcp" (OuterVolumeSpecName: "kube-api-access-rpzcp") pod "346142b6-2a63-4610-b1d0-bfebac8c7c61" (UID: "346142b6-2a63-4610-b1d0-bfebac8c7c61"). InnerVolumeSpecName "kube-api-access-rpzcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:33:07 crc kubenswrapper[4777]: I1124 17:33:07.711785 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/346142b6-2a63-4610-b1d0-bfebac8c7c61-inventory" (OuterVolumeSpecName: "inventory") pod "346142b6-2a63-4610-b1d0-bfebac8c7c61" (UID: "346142b6-2a63-4610-b1d0-bfebac8c7c61"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:33:07 crc kubenswrapper[4777]: I1124 17:33:07.712684 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/346142b6-2a63-4610-b1d0-bfebac8c7c61-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "346142b6-2a63-4610-b1d0-bfebac8c7c61" (UID: "346142b6-2a63-4610-b1d0-bfebac8c7c61"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:33:07 crc kubenswrapper[4777]: I1124 17:33:07.762468 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/346142b6-2a63-4610-b1d0-bfebac8c7c61-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:33:07 crc kubenswrapper[4777]: I1124 17:33:07.762737 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/346142b6-2a63-4610-b1d0-bfebac8c7c61-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:33:07 crc kubenswrapper[4777]: I1124 17:33:07.762822 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpzcp\" (UniqueName: \"kubernetes.io/projected/346142b6-2a63-4610-b1d0-bfebac8c7c61-kube-api-access-rpzcp\") on node \"crc\" DevicePath \"\"" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.017337 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" event={"ID":"346142b6-2a63-4610-b1d0-bfebac8c7c61","Type":"ContainerDied","Data":"a128a35d66debf0eecc8eeca9a7eb8d8203fd31f0f40b8e25bb1ada099d7ff8b"} Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.017715 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a128a35d66debf0eecc8eeca9a7eb8d8203fd31f0f40b8e25bb1ada099d7ff8b" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.017392 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rvlp6" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.113172 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f"] Nov 24 17:33:08 crc kubenswrapper[4777]: E1124 17:33:08.113669 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4825d4-adb7-4f1b-b51b-fe860ad98e3a" containerName="extract-utilities" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.113694 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4825d4-adb7-4f1b-b51b-fe860ad98e3a" containerName="extract-utilities" Nov 24 17:33:08 crc kubenswrapper[4777]: E1124 17:33:08.113744 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4825d4-adb7-4f1b-b51b-fe860ad98e3a" containerName="extract-content" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.113754 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4825d4-adb7-4f1b-b51b-fe860ad98e3a" containerName="extract-content" Nov 24 17:33:08 crc kubenswrapper[4777]: E1124 17:33:08.113771 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="346142b6-2a63-4610-b1d0-bfebac8c7c61" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.113780 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="346142b6-2a63-4610-b1d0-bfebac8c7c61" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:33:08 crc kubenswrapper[4777]: E1124 17:33:08.113794 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4825d4-adb7-4f1b-b51b-fe860ad98e3a" containerName="registry-server" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.113802 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4825d4-adb7-4f1b-b51b-fe860ad98e3a" containerName="registry-server" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.114147 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a4825d4-adb7-4f1b-b51b-fe860ad98e3a" containerName="registry-server" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.114170 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="346142b6-2a63-4610-b1d0-bfebac8c7c61" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.114958 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.117483 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.117705 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.124682 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.125594 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.130500 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f"] Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.171139 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg7g4\" (UniqueName: \"kubernetes.io/projected/b21ec255-1398-4613-b461-29fd5ce00ab1-kube-api-access-dg7g4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qp66f\" (UID: \"b21ec255-1398-4613-b461-29fd5ce00ab1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.171259 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b21ec255-1398-4613-b461-29fd5ce00ab1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qp66f\" (UID: \"b21ec255-1398-4613-b461-29fd5ce00ab1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.171284 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b21ec255-1398-4613-b461-29fd5ce00ab1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qp66f\" (UID: \"b21ec255-1398-4613-b461-29fd5ce00ab1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.273464 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b21ec255-1398-4613-b461-29fd5ce00ab1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qp66f\" (UID: \"b21ec255-1398-4613-b461-29fd5ce00ab1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.273515 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b21ec255-1398-4613-b461-29fd5ce00ab1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qp66f\" (UID: \"b21ec255-1398-4613-b461-29fd5ce00ab1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.273936 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg7g4\" (UniqueName: \"kubernetes.io/projected/b21ec255-1398-4613-b461-29fd5ce00ab1-kube-api-access-dg7g4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qp66f\" (UID: \"b21ec255-1398-4613-b461-29fd5ce00ab1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.280114 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b21ec255-1398-4613-b461-29fd5ce00ab1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qp66f\" (UID: \"b21ec255-1398-4613-b461-29fd5ce00ab1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.280490 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b21ec255-1398-4613-b461-29fd5ce00ab1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qp66f\" (UID: \"b21ec255-1398-4613-b461-29fd5ce00ab1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.292569 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg7g4\" (UniqueName: \"kubernetes.io/projected/b21ec255-1398-4613-b461-29fd5ce00ab1-kube-api-access-dg7g4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qp66f\" (UID: \"b21ec255-1398-4613-b461-29fd5ce00ab1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.433840 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" Nov 24 17:33:08 crc kubenswrapper[4777]: I1124 17:33:08.984341 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f"] Nov 24 17:33:09 crc kubenswrapper[4777]: I1124 17:33:09.032555 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" event={"ID":"b21ec255-1398-4613-b461-29fd5ce00ab1","Type":"ContainerStarted","Data":"0f93bb9a5a0b8150a346cf6c0e1e221bf434c79cb8129dbecc34f1af49116f30"} Nov 24 17:33:10 crc kubenswrapper[4777]: I1124 17:33:10.047394 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" event={"ID":"b21ec255-1398-4613-b461-29fd5ce00ab1","Type":"ContainerStarted","Data":"318da10be50e48404f1d3fde6767930fbad53bcd45ee7d2f026693f546f57954"} Nov 24 17:33:10 crc kubenswrapper[4777]: I1124 17:33:10.088410 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" podStartSLOduration=1.6783356710000001 podStartE2EDuration="2.088393318s" podCreationTimestamp="2025-11-24 17:33:08 +0000 UTC" firstStartedPulling="2025-11-24 17:33:08.992860978 +0000 UTC m=+1947.152056037" lastFinishedPulling="2025-11-24 17:33:09.402918635 +0000 UTC m=+1947.562113684" observedRunningTime="2025-11-24 17:33:10.074812199 +0000 UTC m=+1948.234007288" watchObservedRunningTime="2025-11-24 17:33:10.088393318 +0000 UTC m=+1948.247588367" Nov 24 17:33:24 crc kubenswrapper[4777]: I1124 17:33:24.054449 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-vpbfr"] Nov 24 17:33:24 crc kubenswrapper[4777]: I1124 17:33:24.072917 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-vpbfr"] Nov 24 17:33:25 crc kubenswrapper[4777]: I1124 17:33:25.264251 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4366459e-a3b8-4216-9540-547eeaf67972" path="/var/lib/kubelet/pods/4366459e-a3b8-4216-9540-547eeaf67972/volumes" Nov 24 17:33:32 crc kubenswrapper[4777]: I1124 17:33:32.036525 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xfksm"] Nov 24 17:33:32 crc kubenswrapper[4777]: I1124 17:33:32.053446 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xfksm"] Nov 24 17:33:33 crc kubenswrapper[4777]: I1124 17:33:33.256515 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d534624-b898-4268-a432-7f7a575fb15b" path="/var/lib/kubelet/pods/7d534624-b898-4268-a432-7f7a575fb15b/volumes" Nov 24 17:33:44 crc kubenswrapper[4777]: I1124 17:33:44.270462 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:33:44 crc kubenswrapper[4777]: I1124 17:33:44.271483 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:33:47 crc kubenswrapper[4777]: I1124 17:33:47.939810 4777 scope.go:117] "RemoveContainer" containerID="2c0a4fc5e29a49f0e4e2d0e4b5ffd7824db5cf109cd192786408a0ac53f33e7a" Nov 24 17:33:47 crc kubenswrapper[4777]: I1124 17:33:47.986565 4777 scope.go:117] "RemoveContainer" containerID="8c44adf30eb15f68a0635029c88c02d72b8615d819ae6e6ee0d2957bd4d47abd" Nov 24 17:33:48 crc kubenswrapper[4777]: I1124 17:33:48.034234 4777 scope.go:117] "RemoveContainer" containerID="9d28c6485b14552d3905d31afaa413159ddb8d4913b720e081bc801683048732" Nov 24 17:33:57 crc kubenswrapper[4777]: I1124 17:33:57.569418 4777 generic.go:334] "Generic (PLEG): container finished" podID="b21ec255-1398-4613-b461-29fd5ce00ab1" containerID="318da10be50e48404f1d3fde6767930fbad53bcd45ee7d2f026693f546f57954" exitCode=0 Nov 24 17:33:57 crc kubenswrapper[4777]: I1124 17:33:57.569500 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" event={"ID":"b21ec255-1398-4613-b461-29fd5ce00ab1","Type":"ContainerDied","Data":"318da10be50e48404f1d3fde6767930fbad53bcd45ee7d2f026693f546f57954"} Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.153174 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.285740 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dg7g4\" (UniqueName: \"kubernetes.io/projected/b21ec255-1398-4613-b461-29fd5ce00ab1-kube-api-access-dg7g4\") pod \"b21ec255-1398-4613-b461-29fd5ce00ab1\" (UID: \"b21ec255-1398-4613-b461-29fd5ce00ab1\") " Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.285851 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b21ec255-1398-4613-b461-29fd5ce00ab1-ssh-key\") pod \"b21ec255-1398-4613-b461-29fd5ce00ab1\" (UID: \"b21ec255-1398-4613-b461-29fd5ce00ab1\") " Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.285921 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b21ec255-1398-4613-b461-29fd5ce00ab1-inventory\") pod \"b21ec255-1398-4613-b461-29fd5ce00ab1\" (UID: \"b21ec255-1398-4613-b461-29fd5ce00ab1\") " Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.293173 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b21ec255-1398-4613-b461-29fd5ce00ab1-kube-api-access-dg7g4" (OuterVolumeSpecName: "kube-api-access-dg7g4") pod "b21ec255-1398-4613-b461-29fd5ce00ab1" (UID: "b21ec255-1398-4613-b461-29fd5ce00ab1"). InnerVolumeSpecName "kube-api-access-dg7g4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.316019 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b21ec255-1398-4613-b461-29fd5ce00ab1-inventory" (OuterVolumeSpecName: "inventory") pod "b21ec255-1398-4613-b461-29fd5ce00ab1" (UID: "b21ec255-1398-4613-b461-29fd5ce00ab1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.316296 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b21ec255-1398-4613-b461-29fd5ce00ab1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b21ec255-1398-4613-b461-29fd5ce00ab1" (UID: "b21ec255-1398-4613-b461-29fd5ce00ab1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.388261 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b21ec255-1398-4613-b461-29fd5ce00ab1-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.388302 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b21ec255-1398-4613-b461-29fd5ce00ab1-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.388315 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dg7g4\" (UniqueName: \"kubernetes.io/projected/b21ec255-1398-4613-b461-29fd5ce00ab1-kube-api-access-dg7g4\") on node \"crc\" DevicePath \"\"" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.590359 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" event={"ID":"b21ec255-1398-4613-b461-29fd5ce00ab1","Type":"ContainerDied","Data":"0f93bb9a5a0b8150a346cf6c0e1e221bf434c79cb8129dbecc34f1af49116f30"} Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.590399 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qp66f" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.590405 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f93bb9a5a0b8150a346cf6c0e1e221bf434c79cb8129dbecc34f1af49116f30" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.683160 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-q4nrb"] Nov 24 17:33:59 crc kubenswrapper[4777]: E1124 17:33:59.683718 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b21ec255-1398-4613-b461-29fd5ce00ab1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.683746 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="b21ec255-1398-4613-b461-29fd5ce00ab1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.684066 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="b21ec255-1398-4613-b461-29fd5ce00ab1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.685064 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.688626 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.688717 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.689109 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.689119 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.694624 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-q4nrb"] Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.795679 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3be75c8d-7695-403e-b801-d2650e0b331c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-q4nrb\" (UID: \"3be75c8d-7695-403e-b801-d2650e0b331c\") " pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.795943 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slc5c\" (UniqueName: \"kubernetes.io/projected/3be75c8d-7695-403e-b801-d2650e0b331c-kube-api-access-slc5c\") pod \"ssh-known-hosts-edpm-deployment-q4nrb\" (UID: \"3be75c8d-7695-403e-b801-d2650e0b331c\") " pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.796348 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3be75c8d-7695-403e-b801-d2650e0b331c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-q4nrb\" (UID: \"3be75c8d-7695-403e-b801-d2650e0b331c\") " pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.898352 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slc5c\" (UniqueName: \"kubernetes.io/projected/3be75c8d-7695-403e-b801-d2650e0b331c-kube-api-access-slc5c\") pod \"ssh-known-hosts-edpm-deployment-q4nrb\" (UID: \"3be75c8d-7695-403e-b801-d2650e0b331c\") " pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.898523 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3be75c8d-7695-403e-b801-d2650e0b331c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-q4nrb\" (UID: \"3be75c8d-7695-403e-b801-d2650e0b331c\") " pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.898566 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3be75c8d-7695-403e-b801-d2650e0b331c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-q4nrb\" (UID: \"3be75c8d-7695-403e-b801-d2650e0b331c\") " pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.901935 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3be75c8d-7695-403e-b801-d2650e0b331c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-q4nrb\" (UID: \"3be75c8d-7695-403e-b801-d2650e0b331c\") " pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.903071 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3be75c8d-7695-403e-b801-d2650e0b331c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-q4nrb\" (UID: \"3be75c8d-7695-403e-b801-d2650e0b331c\") " pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" Nov 24 17:33:59 crc kubenswrapper[4777]: I1124 17:33:59.930533 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slc5c\" (UniqueName: \"kubernetes.io/projected/3be75c8d-7695-403e-b801-d2650e0b331c-kube-api-access-slc5c\") pod \"ssh-known-hosts-edpm-deployment-q4nrb\" (UID: \"3be75c8d-7695-403e-b801-d2650e0b331c\") " pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" Nov 24 17:34:00 crc kubenswrapper[4777]: I1124 17:34:00.003985 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" Nov 24 17:34:00 crc kubenswrapper[4777]: I1124 17:34:00.537558 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-q4nrb"] Nov 24 17:34:00 crc kubenswrapper[4777]: I1124 17:34:00.599954 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" event={"ID":"3be75c8d-7695-403e-b801-d2650e0b331c","Type":"ContainerStarted","Data":"92e439ae42a1ac750ee90f3571fab803def7896fcdc92940f91a009cae813586"} Nov 24 17:34:01 crc kubenswrapper[4777]: I1124 17:34:01.614203 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" event={"ID":"3be75c8d-7695-403e-b801-d2650e0b331c","Type":"ContainerStarted","Data":"ef2d28641c335de3bd717c8f719e6f4343db2fb56484539d8ea6466681397eb5"} Nov 24 17:34:01 crc kubenswrapper[4777]: I1124 17:34:01.639446 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" podStartSLOduration=2.205914115 podStartE2EDuration="2.639426935s" podCreationTimestamp="2025-11-24 17:33:59 +0000 UTC" firstStartedPulling="2025-11-24 17:34:00.548093105 +0000 UTC m=+1998.707288144" lastFinishedPulling="2025-11-24 17:34:00.981605915 +0000 UTC m=+1999.140800964" observedRunningTime="2025-11-24 17:34:01.635711819 +0000 UTC m=+1999.794906878" watchObservedRunningTime="2025-11-24 17:34:01.639426935 +0000 UTC m=+1999.798621984" Nov 24 17:34:08 crc kubenswrapper[4777]: I1124 17:34:08.701374 4777 generic.go:334] "Generic (PLEG): container finished" podID="3be75c8d-7695-403e-b801-d2650e0b331c" containerID="ef2d28641c335de3bd717c8f719e6f4343db2fb56484539d8ea6466681397eb5" exitCode=0 Nov 24 17:34:08 crc kubenswrapper[4777]: I1124 17:34:08.701476 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" event={"ID":"3be75c8d-7695-403e-b801-d2650e0b331c","Type":"ContainerDied","Data":"ef2d28641c335de3bd717c8f719e6f4343db2fb56484539d8ea6466681397eb5"} Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.299531 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.433793 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3be75c8d-7695-403e-b801-d2650e0b331c-inventory-0\") pod \"3be75c8d-7695-403e-b801-d2650e0b331c\" (UID: \"3be75c8d-7695-403e-b801-d2650e0b331c\") " Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.434016 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3be75c8d-7695-403e-b801-d2650e0b331c-ssh-key-openstack-edpm-ipam\") pod \"3be75c8d-7695-403e-b801-d2650e0b331c\" (UID: \"3be75c8d-7695-403e-b801-d2650e0b331c\") " Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.434120 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slc5c\" (UniqueName: \"kubernetes.io/projected/3be75c8d-7695-403e-b801-d2650e0b331c-kube-api-access-slc5c\") pod \"3be75c8d-7695-403e-b801-d2650e0b331c\" (UID: \"3be75c8d-7695-403e-b801-d2650e0b331c\") " Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.439187 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3be75c8d-7695-403e-b801-d2650e0b331c-kube-api-access-slc5c" (OuterVolumeSpecName: "kube-api-access-slc5c") pod "3be75c8d-7695-403e-b801-d2650e0b331c" (UID: "3be75c8d-7695-403e-b801-d2650e0b331c"). InnerVolumeSpecName "kube-api-access-slc5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.463617 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3be75c8d-7695-403e-b801-d2650e0b331c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3be75c8d-7695-403e-b801-d2650e0b331c" (UID: "3be75c8d-7695-403e-b801-d2650e0b331c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.465573 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3be75c8d-7695-403e-b801-d2650e0b331c-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "3be75c8d-7695-403e-b801-d2650e0b331c" (UID: "3be75c8d-7695-403e-b801-d2650e0b331c"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.536932 4777 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3be75c8d-7695-403e-b801-d2650e0b331c-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.536989 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3be75c8d-7695-403e-b801-d2650e0b331c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.537008 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slc5c\" (UniqueName: \"kubernetes.io/projected/3be75c8d-7695-403e-b801-d2650e0b331c-kube-api-access-slc5c\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.732710 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" event={"ID":"3be75c8d-7695-403e-b801-d2650e0b331c","Type":"ContainerDied","Data":"92e439ae42a1ac750ee90f3571fab803def7896fcdc92940f91a009cae813586"} Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.732776 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92e439ae42a1ac750ee90f3571fab803def7896fcdc92940f91a009cae813586" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.732843 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-q4nrb" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.809366 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565"] Nov 24 17:34:10 crc kubenswrapper[4777]: E1124 17:34:10.809888 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3be75c8d-7695-403e-b801-d2650e0b331c" containerName="ssh-known-hosts-edpm-deployment" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.809903 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="3be75c8d-7695-403e-b801-d2650e0b331c" containerName="ssh-known-hosts-edpm-deployment" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.810137 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="3be75c8d-7695-403e-b801-d2650e0b331c" containerName="ssh-known-hosts-edpm-deployment" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.810913 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.812950 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.813133 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.813573 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.813708 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.817717 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565"] Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.948621 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkwfs\" (UniqueName: \"kubernetes.io/projected/a57c6a80-0881-430c-a6bb-beea385e0417-kube-api-access-lkwfs\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kq565\" (UID: \"a57c6a80-0881-430c-a6bb-beea385e0417\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.948815 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a57c6a80-0881-430c-a6bb-beea385e0417-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kq565\" (UID: \"a57c6a80-0881-430c-a6bb-beea385e0417\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" Nov 24 17:34:10 crc kubenswrapper[4777]: I1124 17:34:10.948870 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a57c6a80-0881-430c-a6bb-beea385e0417-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kq565\" (UID: \"a57c6a80-0881-430c-a6bb-beea385e0417\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" Nov 24 17:34:11 crc kubenswrapper[4777]: I1124 17:34:11.050647 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a57c6a80-0881-430c-a6bb-beea385e0417-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kq565\" (UID: \"a57c6a80-0881-430c-a6bb-beea385e0417\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" Nov 24 17:34:11 crc kubenswrapper[4777]: I1124 17:34:11.050712 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkwfs\" (UniqueName: \"kubernetes.io/projected/a57c6a80-0881-430c-a6bb-beea385e0417-kube-api-access-lkwfs\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kq565\" (UID: \"a57c6a80-0881-430c-a6bb-beea385e0417\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" Nov 24 17:34:11 crc kubenswrapper[4777]: I1124 17:34:11.050840 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a57c6a80-0881-430c-a6bb-beea385e0417-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kq565\" (UID: \"a57c6a80-0881-430c-a6bb-beea385e0417\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" Nov 24 17:34:11 crc kubenswrapper[4777]: I1124 17:34:11.055125 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a57c6a80-0881-430c-a6bb-beea385e0417-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kq565\" (UID: \"a57c6a80-0881-430c-a6bb-beea385e0417\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" Nov 24 17:34:11 crc kubenswrapper[4777]: I1124 17:34:11.068706 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a57c6a80-0881-430c-a6bb-beea385e0417-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kq565\" (UID: \"a57c6a80-0881-430c-a6bb-beea385e0417\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" Nov 24 17:34:11 crc kubenswrapper[4777]: I1124 17:34:11.071043 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-rjw7h"] Nov 24 17:34:11 crc kubenswrapper[4777]: I1124 17:34:11.076718 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkwfs\" (UniqueName: \"kubernetes.io/projected/a57c6a80-0881-430c-a6bb-beea385e0417-kube-api-access-lkwfs\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kq565\" (UID: \"a57c6a80-0881-430c-a6bb-beea385e0417\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" Nov 24 17:34:11 crc kubenswrapper[4777]: I1124 17:34:11.078671 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-rjw7h"] Nov 24 17:34:11 crc kubenswrapper[4777]: I1124 17:34:11.126801 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" Nov 24 17:34:11 crc kubenswrapper[4777]: I1124 17:34:11.270684 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c1082e3-ae04-40f7-83ae-f7b7299b26b8" path="/var/lib/kubelet/pods/2c1082e3-ae04-40f7-83ae-f7b7299b26b8/volumes" Nov 24 17:34:11 crc kubenswrapper[4777]: I1124 17:34:11.645626 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565"] Nov 24 17:34:11 crc kubenswrapper[4777]: I1124 17:34:11.746257 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" event={"ID":"a57c6a80-0881-430c-a6bb-beea385e0417","Type":"ContainerStarted","Data":"2533ccea6c4011625acac86159bd4565ee6de1a0e5a787e5b0d9e99fb0cf8289"} Nov 24 17:34:12 crc kubenswrapper[4777]: I1124 17:34:12.764516 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" event={"ID":"a57c6a80-0881-430c-a6bb-beea385e0417","Type":"ContainerStarted","Data":"c45a360da0403698d181ccb5b5a5e70b03a154018e5d8d9becda25cb867a9dba"} Nov 24 17:34:12 crc kubenswrapper[4777]: I1124 17:34:12.780421 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" podStartSLOduration=2.129447126 podStartE2EDuration="2.78040114s" podCreationTimestamp="2025-11-24 17:34:10 +0000 UTC" firstStartedPulling="2025-11-24 17:34:11.640805648 +0000 UTC m=+2009.800000697" lastFinishedPulling="2025-11-24 17:34:12.291759622 +0000 UTC m=+2010.450954711" observedRunningTime="2025-11-24 17:34:12.780016689 +0000 UTC m=+2010.939211738" watchObservedRunningTime="2025-11-24 17:34:12.78040114 +0000 UTC m=+2010.939596179" Nov 24 17:34:14 crc kubenswrapper[4777]: I1124 17:34:14.270582 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:34:14 crc kubenswrapper[4777]: I1124 17:34:14.271016 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:34:20 crc kubenswrapper[4777]: I1124 17:34:20.842412 4777 generic.go:334] "Generic (PLEG): container finished" podID="a57c6a80-0881-430c-a6bb-beea385e0417" containerID="c45a360da0403698d181ccb5b5a5e70b03a154018e5d8d9becda25cb867a9dba" exitCode=0 Nov 24 17:34:20 crc kubenswrapper[4777]: I1124 17:34:20.842507 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" event={"ID":"a57c6a80-0881-430c-a6bb-beea385e0417","Type":"ContainerDied","Data":"c45a360da0403698d181ccb5b5a5e70b03a154018e5d8d9becda25cb867a9dba"} Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.338035 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.494011 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a57c6a80-0881-430c-a6bb-beea385e0417-ssh-key\") pod \"a57c6a80-0881-430c-a6bb-beea385e0417\" (UID: \"a57c6a80-0881-430c-a6bb-beea385e0417\") " Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.494118 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkwfs\" (UniqueName: \"kubernetes.io/projected/a57c6a80-0881-430c-a6bb-beea385e0417-kube-api-access-lkwfs\") pod \"a57c6a80-0881-430c-a6bb-beea385e0417\" (UID: \"a57c6a80-0881-430c-a6bb-beea385e0417\") " Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.494137 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a57c6a80-0881-430c-a6bb-beea385e0417-inventory\") pod \"a57c6a80-0881-430c-a6bb-beea385e0417\" (UID: \"a57c6a80-0881-430c-a6bb-beea385e0417\") " Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.500344 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a57c6a80-0881-430c-a6bb-beea385e0417-kube-api-access-lkwfs" (OuterVolumeSpecName: "kube-api-access-lkwfs") pod "a57c6a80-0881-430c-a6bb-beea385e0417" (UID: "a57c6a80-0881-430c-a6bb-beea385e0417"). InnerVolumeSpecName "kube-api-access-lkwfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.526641 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a57c6a80-0881-430c-a6bb-beea385e0417-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a57c6a80-0881-430c-a6bb-beea385e0417" (UID: "a57c6a80-0881-430c-a6bb-beea385e0417"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.557558 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a57c6a80-0881-430c-a6bb-beea385e0417-inventory" (OuterVolumeSpecName: "inventory") pod "a57c6a80-0881-430c-a6bb-beea385e0417" (UID: "a57c6a80-0881-430c-a6bb-beea385e0417"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.596747 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a57c6a80-0881-430c-a6bb-beea385e0417-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.596793 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkwfs\" (UniqueName: \"kubernetes.io/projected/a57c6a80-0881-430c-a6bb-beea385e0417-kube-api-access-lkwfs\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.596814 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a57c6a80-0881-430c-a6bb-beea385e0417-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.865474 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" event={"ID":"a57c6a80-0881-430c-a6bb-beea385e0417","Type":"ContainerDied","Data":"2533ccea6c4011625acac86159bd4565ee6de1a0e5a787e5b0d9e99fb0cf8289"} Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.865763 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2533ccea6c4011625acac86159bd4565ee6de1a0e5a787e5b0d9e99fb0cf8289" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.865535 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kq565" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.948930 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6"] Nov 24 17:34:22 crc kubenswrapper[4777]: E1124 17:34:22.949556 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a57c6a80-0881-430c-a6bb-beea385e0417" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.949581 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="a57c6a80-0881-430c-a6bb-beea385e0417" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.949862 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="a57c6a80-0881-430c-a6bb-beea385e0417" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.950920 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.955183 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.955195 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.955230 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.959292 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:34:22 crc kubenswrapper[4777]: I1124 17:34:22.965311 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6"] Nov 24 17:34:23 crc kubenswrapper[4777]: I1124 17:34:23.106747 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/902e59e0-e5a3-4f96-953e-ec961dbe53d9-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6\" (UID: \"902e59e0-e5a3-4f96-953e-ec961dbe53d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" Nov 24 17:34:23 crc kubenswrapper[4777]: I1124 17:34:23.106953 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/902e59e0-e5a3-4f96-953e-ec961dbe53d9-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6\" (UID: \"902e59e0-e5a3-4f96-953e-ec961dbe53d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" Nov 24 17:34:23 crc kubenswrapper[4777]: I1124 17:34:23.107250 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z747g\" (UniqueName: \"kubernetes.io/projected/902e59e0-e5a3-4f96-953e-ec961dbe53d9-kube-api-access-z747g\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6\" (UID: \"902e59e0-e5a3-4f96-953e-ec961dbe53d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" Nov 24 17:34:23 crc kubenswrapper[4777]: I1124 17:34:23.209309 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/902e59e0-e5a3-4f96-953e-ec961dbe53d9-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6\" (UID: \"902e59e0-e5a3-4f96-953e-ec961dbe53d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" Nov 24 17:34:23 crc kubenswrapper[4777]: I1124 17:34:23.209411 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/902e59e0-e5a3-4f96-953e-ec961dbe53d9-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6\" (UID: \"902e59e0-e5a3-4f96-953e-ec961dbe53d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" Nov 24 17:34:23 crc kubenswrapper[4777]: I1124 17:34:23.210244 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z747g\" (UniqueName: \"kubernetes.io/projected/902e59e0-e5a3-4f96-953e-ec961dbe53d9-kube-api-access-z747g\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6\" (UID: \"902e59e0-e5a3-4f96-953e-ec961dbe53d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" Nov 24 17:34:23 crc kubenswrapper[4777]: I1124 17:34:23.215187 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/902e59e0-e5a3-4f96-953e-ec961dbe53d9-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6\" (UID: \"902e59e0-e5a3-4f96-953e-ec961dbe53d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" Nov 24 17:34:23 crc kubenswrapper[4777]: I1124 17:34:23.215603 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/902e59e0-e5a3-4f96-953e-ec961dbe53d9-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6\" (UID: \"902e59e0-e5a3-4f96-953e-ec961dbe53d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" Nov 24 17:34:23 crc kubenswrapper[4777]: I1124 17:34:23.226740 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z747g\" (UniqueName: \"kubernetes.io/projected/902e59e0-e5a3-4f96-953e-ec961dbe53d9-kube-api-access-z747g\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6\" (UID: \"902e59e0-e5a3-4f96-953e-ec961dbe53d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" Nov 24 17:34:23 crc kubenswrapper[4777]: I1124 17:34:23.279612 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" Nov 24 17:34:23 crc kubenswrapper[4777]: W1124 17:34:23.804683 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod902e59e0_e5a3_4f96_953e_ec961dbe53d9.slice/crio-9ccdafdfb88c25c475bb7dbcbca6e2f6339832b02976e6602848cc6c3ae6ba21 WatchSource:0}: Error finding container 9ccdafdfb88c25c475bb7dbcbca6e2f6339832b02976e6602848cc6c3ae6ba21: Status 404 returned error can't find the container with id 9ccdafdfb88c25c475bb7dbcbca6e2f6339832b02976e6602848cc6c3ae6ba21 Nov 24 17:34:23 crc kubenswrapper[4777]: I1124 17:34:23.805107 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6"] Nov 24 17:34:23 crc kubenswrapper[4777]: I1124 17:34:23.876420 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" event={"ID":"902e59e0-e5a3-4f96-953e-ec961dbe53d9","Type":"ContainerStarted","Data":"9ccdafdfb88c25c475bb7dbcbca6e2f6339832b02976e6602848cc6c3ae6ba21"} Nov 24 17:34:24 crc kubenswrapper[4777]: I1124 17:34:24.887458 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" event={"ID":"902e59e0-e5a3-4f96-953e-ec961dbe53d9","Type":"ContainerStarted","Data":"7a05cbbd53e9edcf2922587648904e478e79ad31bf4d2aa433f4d7e0e06d2f0c"} Nov 24 17:34:24 crc kubenswrapper[4777]: I1124 17:34:24.911628 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" podStartSLOduration=2.30407049 podStartE2EDuration="2.911604469s" podCreationTimestamp="2025-11-24 17:34:22 +0000 UTC" firstStartedPulling="2025-11-24 17:34:23.807617558 +0000 UTC m=+2021.966812597" lastFinishedPulling="2025-11-24 17:34:24.415151527 +0000 UTC m=+2022.574346576" observedRunningTime="2025-11-24 17:34:24.902501828 +0000 UTC m=+2023.061696877" watchObservedRunningTime="2025-11-24 17:34:24.911604469 +0000 UTC m=+2023.070799528" Nov 24 17:34:34 crc kubenswrapper[4777]: I1124 17:34:34.997510 4777 generic.go:334] "Generic (PLEG): container finished" podID="902e59e0-e5a3-4f96-953e-ec961dbe53d9" containerID="7a05cbbd53e9edcf2922587648904e478e79ad31bf4d2aa433f4d7e0e06d2f0c" exitCode=0 Nov 24 17:34:34 crc kubenswrapper[4777]: I1124 17:34:34.998091 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" event={"ID":"902e59e0-e5a3-4f96-953e-ec961dbe53d9","Type":"ContainerDied","Data":"7a05cbbd53e9edcf2922587648904e478e79ad31bf4d2aa433f4d7e0e06d2f0c"} Nov 24 17:34:36 crc kubenswrapper[4777]: I1124 17:34:36.595324 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" Nov 24 17:34:36 crc kubenswrapper[4777]: I1124 17:34:36.742615 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/902e59e0-e5a3-4f96-953e-ec961dbe53d9-inventory\") pod \"902e59e0-e5a3-4f96-953e-ec961dbe53d9\" (UID: \"902e59e0-e5a3-4f96-953e-ec961dbe53d9\") " Nov 24 17:34:36 crc kubenswrapper[4777]: I1124 17:34:36.742686 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z747g\" (UniqueName: \"kubernetes.io/projected/902e59e0-e5a3-4f96-953e-ec961dbe53d9-kube-api-access-z747g\") pod \"902e59e0-e5a3-4f96-953e-ec961dbe53d9\" (UID: \"902e59e0-e5a3-4f96-953e-ec961dbe53d9\") " Nov 24 17:34:36 crc kubenswrapper[4777]: I1124 17:34:36.742996 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/902e59e0-e5a3-4f96-953e-ec961dbe53d9-ssh-key\") pod \"902e59e0-e5a3-4f96-953e-ec961dbe53d9\" (UID: \"902e59e0-e5a3-4f96-953e-ec961dbe53d9\") " Nov 24 17:34:36 crc kubenswrapper[4777]: I1124 17:34:36.752067 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/902e59e0-e5a3-4f96-953e-ec961dbe53d9-kube-api-access-z747g" (OuterVolumeSpecName: "kube-api-access-z747g") pod "902e59e0-e5a3-4f96-953e-ec961dbe53d9" (UID: "902e59e0-e5a3-4f96-953e-ec961dbe53d9"). InnerVolumeSpecName "kube-api-access-z747g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:34:36 crc kubenswrapper[4777]: I1124 17:34:36.779944 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902e59e0-e5a3-4f96-953e-ec961dbe53d9-inventory" (OuterVolumeSpecName: "inventory") pod "902e59e0-e5a3-4f96-953e-ec961dbe53d9" (UID: "902e59e0-e5a3-4f96-953e-ec961dbe53d9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:34:36 crc kubenswrapper[4777]: I1124 17:34:36.792665 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902e59e0-e5a3-4f96-953e-ec961dbe53d9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "902e59e0-e5a3-4f96-953e-ec961dbe53d9" (UID: "902e59e0-e5a3-4f96-953e-ec961dbe53d9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:34:36 crc kubenswrapper[4777]: I1124 17:34:36.846052 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/902e59e0-e5a3-4f96-953e-ec961dbe53d9-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:36 crc kubenswrapper[4777]: I1124 17:34:36.846119 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/902e59e0-e5a3-4f96-953e-ec961dbe53d9-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:36 crc kubenswrapper[4777]: I1124 17:34:36.846137 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z747g\" (UniqueName: \"kubernetes.io/projected/902e59e0-e5a3-4f96-953e-ec961dbe53d9-kube-api-access-z747g\") on node \"crc\" DevicePath \"\"" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.024580 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" event={"ID":"902e59e0-e5a3-4f96-953e-ec961dbe53d9","Type":"ContainerDied","Data":"9ccdafdfb88c25c475bb7dbcbca6e2f6339832b02976e6602848cc6c3ae6ba21"} Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.025023 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ccdafdfb88c25c475bb7dbcbca6e2f6339832b02976e6602848cc6c3ae6ba21" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.025251 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.113481 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw"] Nov 24 17:34:37 crc kubenswrapper[4777]: E1124 17:34:37.114275 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="902e59e0-e5a3-4f96-953e-ec961dbe53d9" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.114294 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="902e59e0-e5a3-4f96-953e-ec961dbe53d9" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.114508 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="902e59e0-e5a3-4f96-953e-ec961dbe53d9" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.115396 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.122393 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.122533 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.122706 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.123180 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.124267 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.124614 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.124886 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.125143 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.141911 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw"] Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.256958 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.257021 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.257044 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.257082 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.257254 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.257314 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.257354 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.257375 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.257432 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.257466 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.257509 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fzhm\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-kube-api-access-9fzhm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.257544 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.257669 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.257705 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.359384 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.359444 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.359465 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.359511 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.359552 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.359574 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.359599 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.359617 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.359645 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.359666 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.359691 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fzhm\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-kube-api-access-9fzhm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.359716 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.359772 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.359792 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.363263 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.363322 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.363344 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.363683 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.364580 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.364625 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.374445 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.374539 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.374689 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.375473 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.375785 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.378075 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fzhm\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-kube-api-access-9fzhm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.378157 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.378481 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:37 crc kubenswrapper[4777]: I1124 17:34:37.435274 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:34:38 crc kubenswrapper[4777]: I1124 17:34:38.064281 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw"] Nov 24 17:34:38 crc kubenswrapper[4777]: W1124 17:34:38.065018 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a98c33f_1b9f_4419_8dd3_ea3ba5695980.slice/crio-ccfcbc4f48852593708eb4dced3af42c4eef445862431105500aea89876e411f WatchSource:0}: Error finding container ccfcbc4f48852593708eb4dced3af42c4eef445862431105500aea89876e411f: Status 404 returned error can't find the container with id ccfcbc4f48852593708eb4dced3af42c4eef445862431105500aea89876e411f Nov 24 17:34:39 crc kubenswrapper[4777]: I1124 17:34:39.047894 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" event={"ID":"0a98c33f-1b9f-4419-8dd3-ea3ba5695980","Type":"ContainerStarted","Data":"ccfcbc4f48852593708eb4dced3af42c4eef445862431105500aea89876e411f"} Nov 24 17:34:41 crc kubenswrapper[4777]: I1124 17:34:41.074923 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" event={"ID":"0a98c33f-1b9f-4419-8dd3-ea3ba5695980","Type":"ContainerStarted","Data":"d17c4d31802e90ea57fa3b0e7b2795e4d1417fcab70f34e0e934113880e3025a"} Nov 24 17:34:44 crc kubenswrapper[4777]: I1124 17:34:44.270231 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:34:44 crc kubenswrapper[4777]: I1124 17:34:44.270951 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:34:44 crc kubenswrapper[4777]: I1124 17:34:44.271045 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:34:44 crc kubenswrapper[4777]: I1124 17:34:44.272688 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a52a61b464393ded65c18d60855e7e57f0df7e2cf82fdbb47caf6735df256f42"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:34:44 crc kubenswrapper[4777]: I1124 17:34:44.272815 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://a52a61b464393ded65c18d60855e7e57f0df7e2cf82fdbb47caf6735df256f42" gracePeriod=600 Nov 24 17:34:45 crc kubenswrapper[4777]: I1124 17:34:45.116295 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="a52a61b464393ded65c18d60855e7e57f0df7e2cf82fdbb47caf6735df256f42" exitCode=0 Nov 24 17:34:45 crc kubenswrapper[4777]: I1124 17:34:45.116476 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"a52a61b464393ded65c18d60855e7e57f0df7e2cf82fdbb47caf6735df256f42"} Nov 24 17:34:45 crc kubenswrapper[4777]: I1124 17:34:45.116768 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d"} Nov 24 17:34:45 crc kubenswrapper[4777]: I1124 17:34:45.116792 4777 scope.go:117] "RemoveContainer" containerID="f1caf4744469bd93e0d67326807bb40e7ffc58dc8fabc282356d8ad225b7a5e4" Nov 24 17:34:45 crc kubenswrapper[4777]: I1124 17:34:45.141554 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" podStartSLOduration=6.269431569 podStartE2EDuration="8.141535148s" podCreationTimestamp="2025-11-24 17:34:37 +0000 UTC" firstStartedPulling="2025-11-24 17:34:38.069033566 +0000 UTC m=+2036.228228625" lastFinishedPulling="2025-11-24 17:34:39.941137155 +0000 UTC m=+2038.100332204" observedRunningTime="2025-11-24 17:34:41.110086447 +0000 UTC m=+2039.269281586" watchObservedRunningTime="2025-11-24 17:34:45.141535148 +0000 UTC m=+2043.300730197" Nov 24 17:34:48 crc kubenswrapper[4777]: I1124 17:34:48.184530 4777 scope.go:117] "RemoveContainer" containerID="1b45413f849b6a67ffd7d2107e94d990e4106b4558326e9226a3e9c6a4c03f4d" Nov 24 17:34:54 crc kubenswrapper[4777]: I1124 17:34:54.043523 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-sync-9d788"] Nov 24 17:34:54 crc kubenswrapper[4777]: I1124 17:34:54.053600 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-sync-9d788"] Nov 24 17:34:55 crc kubenswrapper[4777]: I1124 17:34:55.260174 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e74119f-b07b-4531-b5fa-5a92a9c6625b" path="/var/lib/kubelet/pods/5e74119f-b07b-4531-b5fa-5a92a9c6625b/volumes" Nov 24 17:35:00 crc kubenswrapper[4777]: I1124 17:35:00.033885 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-storageinit-fgmx8"] Nov 24 17:35:00 crc kubenswrapper[4777]: I1124 17:35:00.043745 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-storageinit-fgmx8"] Nov 24 17:35:01 crc kubenswrapper[4777]: I1124 17:35:01.266687 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e5ed725-352f-46ae-83ea-e3b645936482" path="/var/lib/kubelet/pods/0e5ed725-352f-46ae-83ea-e3b645936482/volumes" Nov 24 17:35:17 crc kubenswrapper[4777]: I1124 17:35:17.490503 4777 generic.go:334] "Generic (PLEG): container finished" podID="0a98c33f-1b9f-4419-8dd3-ea3ba5695980" containerID="d17c4d31802e90ea57fa3b0e7b2795e4d1417fcab70f34e0e934113880e3025a" exitCode=0 Nov 24 17:35:17 crc kubenswrapper[4777]: I1124 17:35:17.490598 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" event={"ID":"0a98c33f-1b9f-4419-8dd3-ea3ba5695980","Type":"ContainerDied","Data":"d17c4d31802e90ea57fa3b0e7b2795e4d1417fcab70f34e0e934113880e3025a"} Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.050345 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.071610 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-libvirt-combined-ca-bundle\") pod \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.071660 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-nova-combined-ca-bundle\") pod \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.071699 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-inventory\") pod \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.071745 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-ssh-key\") pod \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.071795 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.071874 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-telemetry-combined-ca-bundle\") pod \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.071913 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fzhm\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-kube-api-access-9fzhm\") pod \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.071961 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.072012 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-repo-setup-combined-ca-bundle\") pod \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.072046 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-bootstrap-combined-ca-bundle\") pod \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.072073 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-neutron-metadata-combined-ca-bundle\") pod \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.072116 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-ovn-combined-ca-bundle\") pod \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.072142 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-ovn-default-certs-0\") pod \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.072245 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\" (UID: \"0a98c33f-1b9f-4419-8dd3-ea3ba5695980\") " Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.081389 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "0a98c33f-1b9f-4419-8dd3-ea3ba5695980" (UID: "0a98c33f-1b9f-4419-8dd3-ea3ba5695980"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.098768 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "0a98c33f-1b9f-4419-8dd3-ea3ba5695980" (UID: "0a98c33f-1b9f-4419-8dd3-ea3ba5695980"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.098781 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "0a98c33f-1b9f-4419-8dd3-ea3ba5695980" (UID: "0a98c33f-1b9f-4419-8dd3-ea3ba5695980"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.098879 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "0a98c33f-1b9f-4419-8dd3-ea3ba5695980" (UID: "0a98c33f-1b9f-4419-8dd3-ea3ba5695980"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.098899 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "0a98c33f-1b9f-4419-8dd3-ea3ba5695980" (UID: "0a98c33f-1b9f-4419-8dd3-ea3ba5695980"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.098904 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-kube-api-access-9fzhm" (OuterVolumeSpecName: "kube-api-access-9fzhm") pod "0a98c33f-1b9f-4419-8dd3-ea3ba5695980" (UID: "0a98c33f-1b9f-4419-8dd3-ea3ba5695980"). InnerVolumeSpecName "kube-api-access-9fzhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.099026 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "0a98c33f-1b9f-4419-8dd3-ea3ba5695980" (UID: "0a98c33f-1b9f-4419-8dd3-ea3ba5695980"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.099049 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "0a98c33f-1b9f-4419-8dd3-ea3ba5695980" (UID: "0a98c33f-1b9f-4419-8dd3-ea3ba5695980"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.099087 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "0a98c33f-1b9f-4419-8dd3-ea3ba5695980" (UID: "0a98c33f-1b9f-4419-8dd3-ea3ba5695980"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.099127 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "0a98c33f-1b9f-4419-8dd3-ea3ba5695980" (UID: "0a98c33f-1b9f-4419-8dd3-ea3ba5695980"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.099162 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "0a98c33f-1b9f-4419-8dd3-ea3ba5695980" (UID: "0a98c33f-1b9f-4419-8dd3-ea3ba5695980"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.100375 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "0a98c33f-1b9f-4419-8dd3-ea3ba5695980" (UID: "0a98c33f-1b9f-4419-8dd3-ea3ba5695980"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.129376 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-inventory" (OuterVolumeSpecName: "inventory") pod "0a98c33f-1b9f-4419-8dd3-ea3ba5695980" (UID: "0a98c33f-1b9f-4419-8dd3-ea3ba5695980"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.141818 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0a98c33f-1b9f-4419-8dd3-ea3ba5695980" (UID: "0a98c33f-1b9f-4419-8dd3-ea3ba5695980"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.175483 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fzhm\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-kube-api-access-9fzhm\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.175538 4777 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.175554 4777 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.175567 4777 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.175579 4777 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.175596 4777 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.175605 4777 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.175615 4777 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.175626 4777 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.175636 4777 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.175645 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.175657 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.175666 4777 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.175677 4777 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a98c33f-1b9f-4419-8dd3-ea3ba5695980-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.517450 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" event={"ID":"0a98c33f-1b9f-4419-8dd3-ea3ba5695980","Type":"ContainerDied","Data":"ccfcbc4f48852593708eb4dced3af42c4eef445862431105500aea89876e411f"} Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.518050 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccfcbc4f48852593708eb4dced3af42c4eef445862431105500aea89876e411f" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.517518 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.611999 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c"] Nov 24 17:35:19 crc kubenswrapper[4777]: E1124 17:35:19.612529 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a98c33f-1b9f-4419-8dd3-ea3ba5695980" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.612567 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a98c33f-1b9f-4419-8dd3-ea3ba5695980" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.612861 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a98c33f-1b9f-4419-8dd3-ea3ba5695980" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.613829 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.616381 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.616438 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.616387 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.617218 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.617233 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.621917 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c"] Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.688254 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p9fw\" (UniqueName: \"kubernetes.io/projected/9c522429-96d7-44c9-812c-f86726228305-kube-api-access-4p9fw\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkd2c\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.688319 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkd2c\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.688403 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkd2c\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.688460 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9c522429-96d7-44c9-812c-f86726228305-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkd2c\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.688533 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkd2c\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.789891 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p9fw\" (UniqueName: \"kubernetes.io/projected/9c522429-96d7-44c9-812c-f86726228305-kube-api-access-4p9fw\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkd2c\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.790239 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkd2c\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.790371 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkd2c\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.791326 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9c522429-96d7-44c9-812c-f86726228305-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkd2c\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.791405 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkd2c\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.793065 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9c522429-96d7-44c9-812c-f86726228305-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkd2c\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.795480 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkd2c\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.796087 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkd2c\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.796718 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkd2c\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.810621 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p9fw\" (UniqueName: \"kubernetes.io/projected/9c522429-96d7-44c9-812c-f86726228305-kube-api-access-4p9fw\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkd2c\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:19 crc kubenswrapper[4777]: I1124 17:35:19.947096 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:35:20 crc kubenswrapper[4777]: I1124 17:35:20.482402 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c"] Nov 24 17:35:20 crc kubenswrapper[4777]: I1124 17:35:20.534172 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" event={"ID":"9c522429-96d7-44c9-812c-f86726228305","Type":"ContainerStarted","Data":"4caad6c443d78dd20f955704d44832dfc1788f048a3efd35342c27fe949f7f9a"} Nov 24 17:35:21 crc kubenswrapper[4777]: I1124 17:35:21.544616 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" event={"ID":"9c522429-96d7-44c9-812c-f86726228305","Type":"ContainerStarted","Data":"4520cbeea5984c8777ca5fa954f9b01aafed448a4e28a02bdb27483969d6f28f"} Nov 24 17:35:21 crc kubenswrapper[4777]: I1124 17:35:21.561639 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" podStartSLOduration=2.111282873 podStartE2EDuration="2.56161825s" podCreationTimestamp="2025-11-24 17:35:19 +0000 UTC" firstStartedPulling="2025-11-24 17:35:20.480019968 +0000 UTC m=+2078.639215007" lastFinishedPulling="2025-11-24 17:35:20.930355335 +0000 UTC m=+2079.089550384" observedRunningTime="2025-11-24 17:35:21.559021157 +0000 UTC m=+2079.718216226" watchObservedRunningTime="2025-11-24 17:35:21.56161825 +0000 UTC m=+2079.720813289" Nov 24 17:35:34 crc kubenswrapper[4777]: I1124 17:35:34.282776 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f4wvv"] Nov 24 17:35:34 crc kubenswrapper[4777]: I1124 17:35:34.297336 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:34 crc kubenswrapper[4777]: I1124 17:35:34.308929 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f4wvv"] Nov 24 17:35:34 crc kubenswrapper[4777]: I1124 17:35:34.427060 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-utilities\") pod \"redhat-marketplace-f4wvv\" (UID: \"3532f3a2-82b4-4e0d-8989-a773c48d1d1f\") " pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:34 crc kubenswrapper[4777]: I1124 17:35:34.427747 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-catalog-content\") pod \"redhat-marketplace-f4wvv\" (UID: \"3532f3a2-82b4-4e0d-8989-a773c48d1d1f\") " pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:34 crc kubenswrapper[4777]: I1124 17:35:34.427841 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csk2b\" (UniqueName: \"kubernetes.io/projected/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-kube-api-access-csk2b\") pod \"redhat-marketplace-f4wvv\" (UID: \"3532f3a2-82b4-4e0d-8989-a773c48d1d1f\") " pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:34 crc kubenswrapper[4777]: I1124 17:35:34.531829 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-catalog-content\") pod \"redhat-marketplace-f4wvv\" (UID: \"3532f3a2-82b4-4e0d-8989-a773c48d1d1f\") " pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:34 crc kubenswrapper[4777]: I1124 17:35:34.531945 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csk2b\" (UniqueName: \"kubernetes.io/projected/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-kube-api-access-csk2b\") pod \"redhat-marketplace-f4wvv\" (UID: \"3532f3a2-82b4-4e0d-8989-a773c48d1d1f\") " pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:34 crc kubenswrapper[4777]: I1124 17:35:34.532087 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-utilities\") pod \"redhat-marketplace-f4wvv\" (UID: \"3532f3a2-82b4-4e0d-8989-a773c48d1d1f\") " pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:34 crc kubenswrapper[4777]: I1124 17:35:34.532615 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-catalog-content\") pod \"redhat-marketplace-f4wvv\" (UID: \"3532f3a2-82b4-4e0d-8989-a773c48d1d1f\") " pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:34 crc kubenswrapper[4777]: I1124 17:35:34.533475 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-utilities\") pod \"redhat-marketplace-f4wvv\" (UID: \"3532f3a2-82b4-4e0d-8989-a773c48d1d1f\") " pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:34 crc kubenswrapper[4777]: I1124 17:35:34.554873 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csk2b\" (UniqueName: \"kubernetes.io/projected/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-kube-api-access-csk2b\") pod \"redhat-marketplace-f4wvv\" (UID: \"3532f3a2-82b4-4e0d-8989-a773c48d1d1f\") " pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:34 crc kubenswrapper[4777]: I1124 17:35:34.628141 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:35 crc kubenswrapper[4777]: I1124 17:35:35.139050 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f4wvv"] Nov 24 17:35:35 crc kubenswrapper[4777]: W1124 17:35:35.146394 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3532f3a2_82b4_4e0d_8989_a773c48d1d1f.slice/crio-3b1f4f35673ec75e8d44049db0fbd8e7e942f851e84f4e9bea09795a35a4a378 WatchSource:0}: Error finding container 3b1f4f35673ec75e8d44049db0fbd8e7e942f851e84f4e9bea09795a35a4a378: Status 404 returned error can't find the container with id 3b1f4f35673ec75e8d44049db0fbd8e7e942f851e84f4e9bea09795a35a4a378 Nov 24 17:35:35 crc kubenswrapper[4777]: I1124 17:35:35.686858 4777 generic.go:334] "Generic (PLEG): container finished" podID="3532f3a2-82b4-4e0d-8989-a773c48d1d1f" containerID="7d7f9596031dc95f846b0da6fbcbe8f815b2ae590b2a1f9560b99d901fa5d5b0" exitCode=0 Nov 24 17:35:35 crc kubenswrapper[4777]: I1124 17:35:35.687113 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f4wvv" event={"ID":"3532f3a2-82b4-4e0d-8989-a773c48d1d1f","Type":"ContainerDied","Data":"7d7f9596031dc95f846b0da6fbcbe8f815b2ae590b2a1f9560b99d901fa5d5b0"} Nov 24 17:35:35 crc kubenswrapper[4777]: I1124 17:35:35.687205 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f4wvv" event={"ID":"3532f3a2-82b4-4e0d-8989-a773c48d1d1f","Type":"ContainerStarted","Data":"3b1f4f35673ec75e8d44049db0fbd8e7e942f851e84f4e9bea09795a35a4a378"} Nov 24 17:35:37 crc kubenswrapper[4777]: I1124 17:35:37.711275 4777 generic.go:334] "Generic (PLEG): container finished" podID="3532f3a2-82b4-4e0d-8989-a773c48d1d1f" containerID="db4e3c5a3be8ecc0b3478af79b4322e35af443cf88c1c25106e39386d07ef410" exitCode=0 Nov 24 17:35:37 crc kubenswrapper[4777]: I1124 17:35:37.711389 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f4wvv" event={"ID":"3532f3a2-82b4-4e0d-8989-a773c48d1d1f","Type":"ContainerDied","Data":"db4e3c5a3be8ecc0b3478af79b4322e35af443cf88c1c25106e39386d07ef410"} Nov 24 17:35:38 crc kubenswrapper[4777]: I1124 17:35:38.724014 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f4wvv" event={"ID":"3532f3a2-82b4-4e0d-8989-a773c48d1d1f","Type":"ContainerStarted","Data":"216b3730e0bfeb5e90a774b1588f5f0f75357eb05fb36cc3701dc3baf1a9b2e7"} Nov 24 17:35:38 crc kubenswrapper[4777]: I1124 17:35:38.756103 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f4wvv" podStartSLOduration=2.338592699 podStartE2EDuration="4.756083378s" podCreationTimestamp="2025-11-24 17:35:34 +0000 UTC" firstStartedPulling="2025-11-24 17:35:35.68935035 +0000 UTC m=+2093.848545399" lastFinishedPulling="2025-11-24 17:35:38.106840999 +0000 UTC m=+2096.266036078" observedRunningTime="2025-11-24 17:35:38.751078868 +0000 UTC m=+2096.910273917" watchObservedRunningTime="2025-11-24 17:35:38.756083378 +0000 UTC m=+2096.915278427" Nov 24 17:35:44 crc kubenswrapper[4777]: I1124 17:35:44.629207 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:44 crc kubenswrapper[4777]: I1124 17:35:44.630033 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:44 crc kubenswrapper[4777]: I1124 17:35:44.677587 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:44 crc kubenswrapper[4777]: I1124 17:35:44.836002 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:44 crc kubenswrapper[4777]: I1124 17:35:44.920643 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f4wvv"] Nov 24 17:35:46 crc kubenswrapper[4777]: I1124 17:35:46.831713 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f4wvv" podUID="3532f3a2-82b4-4e0d-8989-a773c48d1d1f" containerName="registry-server" containerID="cri-o://216b3730e0bfeb5e90a774b1588f5f0f75357eb05fb36cc3701dc3baf1a9b2e7" gracePeriod=2 Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.439875 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.523541 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csk2b\" (UniqueName: \"kubernetes.io/projected/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-kube-api-access-csk2b\") pod \"3532f3a2-82b4-4e0d-8989-a773c48d1d1f\" (UID: \"3532f3a2-82b4-4e0d-8989-a773c48d1d1f\") " Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.523656 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-utilities\") pod \"3532f3a2-82b4-4e0d-8989-a773c48d1d1f\" (UID: \"3532f3a2-82b4-4e0d-8989-a773c48d1d1f\") " Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.523715 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-catalog-content\") pod \"3532f3a2-82b4-4e0d-8989-a773c48d1d1f\" (UID: \"3532f3a2-82b4-4e0d-8989-a773c48d1d1f\") " Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.524547 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-utilities" (OuterVolumeSpecName: "utilities") pod "3532f3a2-82b4-4e0d-8989-a773c48d1d1f" (UID: "3532f3a2-82b4-4e0d-8989-a773c48d1d1f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.537616 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-kube-api-access-csk2b" (OuterVolumeSpecName: "kube-api-access-csk2b") pod "3532f3a2-82b4-4e0d-8989-a773c48d1d1f" (UID: "3532f3a2-82b4-4e0d-8989-a773c48d1d1f"). InnerVolumeSpecName "kube-api-access-csk2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.542505 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3532f3a2-82b4-4e0d-8989-a773c48d1d1f" (UID: "3532f3a2-82b4-4e0d-8989-a773c48d1d1f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.627378 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csk2b\" (UniqueName: \"kubernetes.io/projected/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-kube-api-access-csk2b\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.627933 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.627951 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3532f3a2-82b4-4e0d-8989-a773c48d1d1f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.853533 4777 generic.go:334] "Generic (PLEG): container finished" podID="3532f3a2-82b4-4e0d-8989-a773c48d1d1f" containerID="216b3730e0bfeb5e90a774b1588f5f0f75357eb05fb36cc3701dc3baf1a9b2e7" exitCode=0 Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.853590 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f4wvv" event={"ID":"3532f3a2-82b4-4e0d-8989-a773c48d1d1f","Type":"ContainerDied","Data":"216b3730e0bfeb5e90a774b1588f5f0f75357eb05fb36cc3701dc3baf1a9b2e7"} Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.853630 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f4wvv" event={"ID":"3532f3a2-82b4-4e0d-8989-a773c48d1d1f","Type":"ContainerDied","Data":"3b1f4f35673ec75e8d44049db0fbd8e7e942f851e84f4e9bea09795a35a4a378"} Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.853657 4777 scope.go:117] "RemoveContainer" containerID="216b3730e0bfeb5e90a774b1588f5f0f75357eb05fb36cc3701dc3baf1a9b2e7" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.853770 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f4wvv" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.885013 4777 scope.go:117] "RemoveContainer" containerID="db4e3c5a3be8ecc0b3478af79b4322e35af443cf88c1c25106e39386d07ef410" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.900067 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f4wvv"] Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.911664 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f4wvv"] Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.917318 4777 scope.go:117] "RemoveContainer" containerID="7d7f9596031dc95f846b0da6fbcbe8f815b2ae590b2a1f9560b99d901fa5d5b0" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.954364 4777 scope.go:117] "RemoveContainer" containerID="216b3730e0bfeb5e90a774b1588f5f0f75357eb05fb36cc3701dc3baf1a9b2e7" Nov 24 17:35:47 crc kubenswrapper[4777]: E1124 17:35:47.954858 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"216b3730e0bfeb5e90a774b1588f5f0f75357eb05fb36cc3701dc3baf1a9b2e7\": container with ID starting with 216b3730e0bfeb5e90a774b1588f5f0f75357eb05fb36cc3701dc3baf1a9b2e7 not found: ID does not exist" containerID="216b3730e0bfeb5e90a774b1588f5f0f75357eb05fb36cc3701dc3baf1a9b2e7" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.954894 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"216b3730e0bfeb5e90a774b1588f5f0f75357eb05fb36cc3701dc3baf1a9b2e7"} err="failed to get container status \"216b3730e0bfeb5e90a774b1588f5f0f75357eb05fb36cc3701dc3baf1a9b2e7\": rpc error: code = NotFound desc = could not find container \"216b3730e0bfeb5e90a774b1588f5f0f75357eb05fb36cc3701dc3baf1a9b2e7\": container with ID starting with 216b3730e0bfeb5e90a774b1588f5f0f75357eb05fb36cc3701dc3baf1a9b2e7 not found: ID does not exist" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.954919 4777 scope.go:117] "RemoveContainer" containerID="db4e3c5a3be8ecc0b3478af79b4322e35af443cf88c1c25106e39386d07ef410" Nov 24 17:35:47 crc kubenswrapper[4777]: E1124 17:35:47.955330 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db4e3c5a3be8ecc0b3478af79b4322e35af443cf88c1c25106e39386d07ef410\": container with ID starting with db4e3c5a3be8ecc0b3478af79b4322e35af443cf88c1c25106e39386d07ef410 not found: ID does not exist" containerID="db4e3c5a3be8ecc0b3478af79b4322e35af443cf88c1c25106e39386d07ef410" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.955362 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4e3c5a3be8ecc0b3478af79b4322e35af443cf88c1c25106e39386d07ef410"} err="failed to get container status \"db4e3c5a3be8ecc0b3478af79b4322e35af443cf88c1c25106e39386d07ef410\": rpc error: code = NotFound desc = could not find container \"db4e3c5a3be8ecc0b3478af79b4322e35af443cf88c1c25106e39386d07ef410\": container with ID starting with db4e3c5a3be8ecc0b3478af79b4322e35af443cf88c1c25106e39386d07ef410 not found: ID does not exist" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.955376 4777 scope.go:117] "RemoveContainer" containerID="7d7f9596031dc95f846b0da6fbcbe8f815b2ae590b2a1f9560b99d901fa5d5b0" Nov 24 17:35:47 crc kubenswrapper[4777]: E1124 17:35:47.955624 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d7f9596031dc95f846b0da6fbcbe8f815b2ae590b2a1f9560b99d901fa5d5b0\": container with ID starting with 7d7f9596031dc95f846b0da6fbcbe8f815b2ae590b2a1f9560b99d901fa5d5b0 not found: ID does not exist" containerID="7d7f9596031dc95f846b0da6fbcbe8f815b2ae590b2a1f9560b99d901fa5d5b0" Nov 24 17:35:47 crc kubenswrapper[4777]: I1124 17:35:47.955648 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d7f9596031dc95f846b0da6fbcbe8f815b2ae590b2a1f9560b99d901fa5d5b0"} err="failed to get container status \"7d7f9596031dc95f846b0da6fbcbe8f815b2ae590b2a1f9560b99d901fa5d5b0\": rpc error: code = NotFound desc = could not find container \"7d7f9596031dc95f846b0da6fbcbe8f815b2ae590b2a1f9560b99d901fa5d5b0\": container with ID starting with 7d7f9596031dc95f846b0da6fbcbe8f815b2ae590b2a1f9560b99d901fa5d5b0 not found: ID does not exist" Nov 24 17:35:48 crc kubenswrapper[4777]: I1124 17:35:48.346539 4777 scope.go:117] "RemoveContainer" containerID="af536934b889d96923c0857686c62be0501f1bd4a996dbc0d04dc714a2a1bdbf" Nov 24 17:35:48 crc kubenswrapper[4777]: I1124 17:35:48.372260 4777 scope.go:117] "RemoveContainer" containerID="828c4ec9b88e85872599f26f0bc8dbf67daad7b0252833a2b91593572ab5062b" Nov 24 17:35:49 crc kubenswrapper[4777]: I1124 17:35:49.258465 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3532f3a2-82b4-4e0d-8989-a773c48d1d1f" path="/var/lib/kubelet/pods/3532f3a2-82b4-4e0d-8989-a773c48d1d1f/volumes" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.172502 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-psgx9"] Nov 24 17:35:52 crc kubenswrapper[4777]: E1124 17:35:52.173406 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3532f3a2-82b4-4e0d-8989-a773c48d1d1f" containerName="extract-content" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.173425 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="3532f3a2-82b4-4e0d-8989-a773c48d1d1f" containerName="extract-content" Nov 24 17:35:52 crc kubenswrapper[4777]: E1124 17:35:52.173464 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3532f3a2-82b4-4e0d-8989-a773c48d1d1f" containerName="extract-utilities" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.173473 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="3532f3a2-82b4-4e0d-8989-a773c48d1d1f" containerName="extract-utilities" Nov 24 17:35:52 crc kubenswrapper[4777]: E1124 17:35:52.173502 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3532f3a2-82b4-4e0d-8989-a773c48d1d1f" containerName="registry-server" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.173510 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="3532f3a2-82b4-4e0d-8989-a773c48d1d1f" containerName="registry-server" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.173772 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="3532f3a2-82b4-4e0d-8989-a773c48d1d1f" containerName="registry-server" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.175600 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.193207 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-psgx9"] Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.224470 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lglv\" (UniqueName: \"kubernetes.io/projected/9e693b21-413a-4ae2-bbfd-8b860359a74e-kube-api-access-5lglv\") pod \"community-operators-psgx9\" (UID: \"9e693b21-413a-4ae2-bbfd-8b860359a74e\") " pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.224574 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e693b21-413a-4ae2-bbfd-8b860359a74e-utilities\") pod \"community-operators-psgx9\" (UID: \"9e693b21-413a-4ae2-bbfd-8b860359a74e\") " pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.224711 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e693b21-413a-4ae2-bbfd-8b860359a74e-catalog-content\") pod \"community-operators-psgx9\" (UID: \"9e693b21-413a-4ae2-bbfd-8b860359a74e\") " pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.326505 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e693b21-413a-4ae2-bbfd-8b860359a74e-catalog-content\") pod \"community-operators-psgx9\" (UID: \"9e693b21-413a-4ae2-bbfd-8b860359a74e\") " pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.326605 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lglv\" (UniqueName: \"kubernetes.io/projected/9e693b21-413a-4ae2-bbfd-8b860359a74e-kube-api-access-5lglv\") pod \"community-operators-psgx9\" (UID: \"9e693b21-413a-4ae2-bbfd-8b860359a74e\") " pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.326713 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e693b21-413a-4ae2-bbfd-8b860359a74e-utilities\") pod \"community-operators-psgx9\" (UID: \"9e693b21-413a-4ae2-bbfd-8b860359a74e\") " pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.327137 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e693b21-413a-4ae2-bbfd-8b860359a74e-catalog-content\") pod \"community-operators-psgx9\" (UID: \"9e693b21-413a-4ae2-bbfd-8b860359a74e\") " pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.327427 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e693b21-413a-4ae2-bbfd-8b860359a74e-utilities\") pod \"community-operators-psgx9\" (UID: \"9e693b21-413a-4ae2-bbfd-8b860359a74e\") " pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.348603 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lglv\" (UniqueName: \"kubernetes.io/projected/9e693b21-413a-4ae2-bbfd-8b860359a74e-kube-api-access-5lglv\") pod \"community-operators-psgx9\" (UID: \"9e693b21-413a-4ae2-bbfd-8b860359a74e\") " pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:35:52 crc kubenswrapper[4777]: I1124 17:35:52.510522 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:35:53 crc kubenswrapper[4777]: I1124 17:35:53.041484 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-psgx9"] Nov 24 17:35:53 crc kubenswrapper[4777]: I1124 17:35:53.939072 4777 generic.go:334] "Generic (PLEG): container finished" podID="9e693b21-413a-4ae2-bbfd-8b860359a74e" containerID="c18ee30da80bfec1bac9bfad40576b7e106385dd38b91c1af3f433d54d978f1e" exitCode=0 Nov 24 17:35:53 crc kubenswrapper[4777]: I1124 17:35:53.939712 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-psgx9" event={"ID":"9e693b21-413a-4ae2-bbfd-8b860359a74e","Type":"ContainerDied","Data":"c18ee30da80bfec1bac9bfad40576b7e106385dd38b91c1af3f433d54d978f1e"} Nov 24 17:35:53 crc kubenswrapper[4777]: I1124 17:35:53.939783 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-psgx9" event={"ID":"9e693b21-413a-4ae2-bbfd-8b860359a74e","Type":"ContainerStarted","Data":"5b7aa8ebd4388f80200d0286ce63da71b501a4dea96538790f18faf3ae80f332"} Nov 24 17:35:54 crc kubenswrapper[4777]: I1124 17:35:54.955786 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-psgx9" event={"ID":"9e693b21-413a-4ae2-bbfd-8b860359a74e","Type":"ContainerStarted","Data":"b64881128b236a913c0fcf2fe1007a774c401224a7e5f8b2953eea307262b5b7"} Nov 24 17:35:56 crc kubenswrapper[4777]: I1124 17:35:56.976530 4777 generic.go:334] "Generic (PLEG): container finished" podID="9e693b21-413a-4ae2-bbfd-8b860359a74e" containerID="b64881128b236a913c0fcf2fe1007a774c401224a7e5f8b2953eea307262b5b7" exitCode=0 Nov 24 17:35:56 crc kubenswrapper[4777]: I1124 17:35:56.976579 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-psgx9" event={"ID":"9e693b21-413a-4ae2-bbfd-8b860359a74e","Type":"ContainerDied","Data":"b64881128b236a913c0fcf2fe1007a774c401224a7e5f8b2953eea307262b5b7"} Nov 24 17:35:57 crc kubenswrapper[4777]: I1124 17:35:57.988700 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-psgx9" event={"ID":"9e693b21-413a-4ae2-bbfd-8b860359a74e","Type":"ContainerStarted","Data":"6909c6b0067875a38fa5d1f71a69177fc2efca20f041a949627bfa8ef5f277da"} Nov 24 17:35:58 crc kubenswrapper[4777]: I1124 17:35:58.024132 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-psgx9" podStartSLOduration=2.593178194 podStartE2EDuration="6.024108773s" podCreationTimestamp="2025-11-24 17:35:52 +0000 UTC" firstStartedPulling="2025-11-24 17:35:53.943638088 +0000 UTC m=+2112.102833147" lastFinishedPulling="2025-11-24 17:35:57.374568667 +0000 UTC m=+2115.533763726" observedRunningTime="2025-11-24 17:35:58.014994207 +0000 UTC m=+2116.174189256" watchObservedRunningTime="2025-11-24 17:35:58.024108773 +0000 UTC m=+2116.183303822" Nov 24 17:36:02 crc kubenswrapper[4777]: I1124 17:36:02.511144 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:36:02 crc kubenswrapper[4777]: I1124 17:36:02.513259 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:36:02 crc kubenswrapper[4777]: I1124 17:36:02.560304 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:36:03 crc kubenswrapper[4777]: I1124 17:36:03.082620 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:36:03 crc kubenswrapper[4777]: I1124 17:36:03.134696 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-psgx9"] Nov 24 17:36:05 crc kubenswrapper[4777]: I1124 17:36:05.053693 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-psgx9" podUID="9e693b21-413a-4ae2-bbfd-8b860359a74e" containerName="registry-server" containerID="cri-o://6909c6b0067875a38fa5d1f71a69177fc2efca20f041a949627bfa8ef5f277da" gracePeriod=2 Nov 24 17:36:05 crc kubenswrapper[4777]: I1124 17:36:05.594678 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:36:05 crc kubenswrapper[4777]: I1124 17:36:05.712497 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e693b21-413a-4ae2-bbfd-8b860359a74e-catalog-content\") pod \"9e693b21-413a-4ae2-bbfd-8b860359a74e\" (UID: \"9e693b21-413a-4ae2-bbfd-8b860359a74e\") " Nov 24 17:36:05 crc kubenswrapper[4777]: I1124 17:36:05.712643 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lglv\" (UniqueName: \"kubernetes.io/projected/9e693b21-413a-4ae2-bbfd-8b860359a74e-kube-api-access-5lglv\") pod \"9e693b21-413a-4ae2-bbfd-8b860359a74e\" (UID: \"9e693b21-413a-4ae2-bbfd-8b860359a74e\") " Nov 24 17:36:05 crc kubenswrapper[4777]: I1124 17:36:05.713176 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e693b21-413a-4ae2-bbfd-8b860359a74e-utilities\") pod \"9e693b21-413a-4ae2-bbfd-8b860359a74e\" (UID: \"9e693b21-413a-4ae2-bbfd-8b860359a74e\") " Nov 24 17:36:05 crc kubenswrapper[4777]: I1124 17:36:05.713901 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e693b21-413a-4ae2-bbfd-8b860359a74e-utilities" (OuterVolumeSpecName: "utilities") pod "9e693b21-413a-4ae2-bbfd-8b860359a74e" (UID: "9e693b21-413a-4ae2-bbfd-8b860359a74e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:36:05 crc kubenswrapper[4777]: I1124 17:36:05.719935 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e693b21-413a-4ae2-bbfd-8b860359a74e-kube-api-access-5lglv" (OuterVolumeSpecName: "kube-api-access-5lglv") pod "9e693b21-413a-4ae2-bbfd-8b860359a74e" (UID: "9e693b21-413a-4ae2-bbfd-8b860359a74e"). InnerVolumeSpecName "kube-api-access-5lglv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:36:05 crc kubenswrapper[4777]: I1124 17:36:05.778516 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e693b21-413a-4ae2-bbfd-8b860359a74e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e693b21-413a-4ae2-bbfd-8b860359a74e" (UID: "9e693b21-413a-4ae2-bbfd-8b860359a74e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:36:05 crc kubenswrapper[4777]: I1124 17:36:05.816145 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lglv\" (UniqueName: \"kubernetes.io/projected/9e693b21-413a-4ae2-bbfd-8b860359a74e-kube-api-access-5lglv\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:05 crc kubenswrapper[4777]: I1124 17:36:05.816182 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e693b21-413a-4ae2-bbfd-8b860359a74e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:05 crc kubenswrapper[4777]: I1124 17:36:05.816191 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e693b21-413a-4ae2-bbfd-8b860359a74e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:06 crc kubenswrapper[4777]: I1124 17:36:06.066039 4777 generic.go:334] "Generic (PLEG): container finished" podID="9e693b21-413a-4ae2-bbfd-8b860359a74e" containerID="6909c6b0067875a38fa5d1f71a69177fc2efca20f041a949627bfa8ef5f277da" exitCode=0 Nov 24 17:36:06 crc kubenswrapper[4777]: I1124 17:36:06.066114 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-psgx9" Nov 24 17:36:06 crc kubenswrapper[4777]: I1124 17:36:06.066101 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-psgx9" event={"ID":"9e693b21-413a-4ae2-bbfd-8b860359a74e","Type":"ContainerDied","Data":"6909c6b0067875a38fa5d1f71a69177fc2efca20f041a949627bfa8ef5f277da"} Nov 24 17:36:06 crc kubenswrapper[4777]: I1124 17:36:06.066515 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-psgx9" event={"ID":"9e693b21-413a-4ae2-bbfd-8b860359a74e","Type":"ContainerDied","Data":"5b7aa8ebd4388f80200d0286ce63da71b501a4dea96538790f18faf3ae80f332"} Nov 24 17:36:06 crc kubenswrapper[4777]: I1124 17:36:06.066540 4777 scope.go:117] "RemoveContainer" containerID="6909c6b0067875a38fa5d1f71a69177fc2efca20f041a949627bfa8ef5f277da" Nov 24 17:36:06 crc kubenswrapper[4777]: I1124 17:36:06.096341 4777 scope.go:117] "RemoveContainer" containerID="b64881128b236a913c0fcf2fe1007a774c401224a7e5f8b2953eea307262b5b7" Nov 24 17:36:06 crc kubenswrapper[4777]: I1124 17:36:06.100775 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-psgx9"] Nov 24 17:36:06 crc kubenswrapper[4777]: I1124 17:36:06.109769 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-psgx9"] Nov 24 17:36:06 crc kubenswrapper[4777]: I1124 17:36:06.123300 4777 scope.go:117] "RemoveContainer" containerID="c18ee30da80bfec1bac9bfad40576b7e106385dd38b91c1af3f433d54d978f1e" Nov 24 17:36:06 crc kubenswrapper[4777]: I1124 17:36:06.184962 4777 scope.go:117] "RemoveContainer" containerID="6909c6b0067875a38fa5d1f71a69177fc2efca20f041a949627bfa8ef5f277da" Nov 24 17:36:06 crc kubenswrapper[4777]: E1124 17:36:06.185530 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6909c6b0067875a38fa5d1f71a69177fc2efca20f041a949627bfa8ef5f277da\": container with ID starting with 6909c6b0067875a38fa5d1f71a69177fc2efca20f041a949627bfa8ef5f277da not found: ID does not exist" containerID="6909c6b0067875a38fa5d1f71a69177fc2efca20f041a949627bfa8ef5f277da" Nov 24 17:36:06 crc kubenswrapper[4777]: I1124 17:36:06.185569 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6909c6b0067875a38fa5d1f71a69177fc2efca20f041a949627bfa8ef5f277da"} err="failed to get container status \"6909c6b0067875a38fa5d1f71a69177fc2efca20f041a949627bfa8ef5f277da\": rpc error: code = NotFound desc = could not find container \"6909c6b0067875a38fa5d1f71a69177fc2efca20f041a949627bfa8ef5f277da\": container with ID starting with 6909c6b0067875a38fa5d1f71a69177fc2efca20f041a949627bfa8ef5f277da not found: ID does not exist" Nov 24 17:36:06 crc kubenswrapper[4777]: I1124 17:36:06.185600 4777 scope.go:117] "RemoveContainer" containerID="b64881128b236a913c0fcf2fe1007a774c401224a7e5f8b2953eea307262b5b7" Nov 24 17:36:06 crc kubenswrapper[4777]: E1124 17:36:06.186329 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b64881128b236a913c0fcf2fe1007a774c401224a7e5f8b2953eea307262b5b7\": container with ID starting with b64881128b236a913c0fcf2fe1007a774c401224a7e5f8b2953eea307262b5b7 not found: ID does not exist" containerID="b64881128b236a913c0fcf2fe1007a774c401224a7e5f8b2953eea307262b5b7" Nov 24 17:36:06 crc kubenswrapper[4777]: I1124 17:36:06.186361 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b64881128b236a913c0fcf2fe1007a774c401224a7e5f8b2953eea307262b5b7"} err="failed to get container status \"b64881128b236a913c0fcf2fe1007a774c401224a7e5f8b2953eea307262b5b7\": rpc error: code = NotFound desc = could not find container \"b64881128b236a913c0fcf2fe1007a774c401224a7e5f8b2953eea307262b5b7\": container with ID starting with b64881128b236a913c0fcf2fe1007a774c401224a7e5f8b2953eea307262b5b7 not found: ID does not exist" Nov 24 17:36:06 crc kubenswrapper[4777]: I1124 17:36:06.186384 4777 scope.go:117] "RemoveContainer" containerID="c18ee30da80bfec1bac9bfad40576b7e106385dd38b91c1af3f433d54d978f1e" Nov 24 17:36:06 crc kubenswrapper[4777]: E1124 17:36:06.186738 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c18ee30da80bfec1bac9bfad40576b7e106385dd38b91c1af3f433d54d978f1e\": container with ID starting with c18ee30da80bfec1bac9bfad40576b7e106385dd38b91c1af3f433d54d978f1e not found: ID does not exist" containerID="c18ee30da80bfec1bac9bfad40576b7e106385dd38b91c1af3f433d54d978f1e" Nov 24 17:36:06 crc kubenswrapper[4777]: I1124 17:36:06.186761 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c18ee30da80bfec1bac9bfad40576b7e106385dd38b91c1af3f433d54d978f1e"} err="failed to get container status \"c18ee30da80bfec1bac9bfad40576b7e106385dd38b91c1af3f433d54d978f1e\": rpc error: code = NotFound desc = could not find container \"c18ee30da80bfec1bac9bfad40576b7e106385dd38b91c1af3f433d54d978f1e\": container with ID starting with c18ee30da80bfec1bac9bfad40576b7e106385dd38b91c1af3f433d54d978f1e not found: ID does not exist" Nov 24 17:36:07 crc kubenswrapper[4777]: I1124 17:36:07.273506 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e693b21-413a-4ae2-bbfd-8b860359a74e" path="/var/lib/kubelet/pods/9e693b21-413a-4ae2-bbfd-8b860359a74e/volumes" Nov 24 17:36:21 crc kubenswrapper[4777]: I1124 17:36:21.286566 4777 generic.go:334] "Generic (PLEG): container finished" podID="9c522429-96d7-44c9-812c-f86726228305" containerID="4520cbeea5984c8777ca5fa954f9b01aafed448a4e28a02bdb27483969d6f28f" exitCode=0 Nov 24 17:36:21 crc kubenswrapper[4777]: I1124 17:36:21.286646 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" event={"ID":"9c522429-96d7-44c9-812c-f86726228305","Type":"ContainerDied","Data":"4520cbeea5984c8777ca5fa954f9b01aafed448a4e28a02bdb27483969d6f28f"} Nov 24 17:36:22 crc kubenswrapper[4777]: I1124 17:36:22.856918 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:36:22 crc kubenswrapper[4777]: I1124 17:36:22.979477 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-ovn-combined-ca-bundle\") pod \"9c522429-96d7-44c9-812c-f86726228305\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " Nov 24 17:36:22 crc kubenswrapper[4777]: I1124 17:36:22.979610 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9c522429-96d7-44c9-812c-f86726228305-ovncontroller-config-0\") pod \"9c522429-96d7-44c9-812c-f86726228305\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " Nov 24 17:36:22 crc kubenswrapper[4777]: I1124 17:36:22.979695 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-ssh-key\") pod \"9c522429-96d7-44c9-812c-f86726228305\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " Nov 24 17:36:22 crc kubenswrapper[4777]: I1124 17:36:22.979726 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4p9fw\" (UniqueName: \"kubernetes.io/projected/9c522429-96d7-44c9-812c-f86726228305-kube-api-access-4p9fw\") pod \"9c522429-96d7-44c9-812c-f86726228305\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " Nov 24 17:36:22 crc kubenswrapper[4777]: I1124 17:36:22.980660 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-inventory\") pod \"9c522429-96d7-44c9-812c-f86726228305\" (UID: \"9c522429-96d7-44c9-812c-f86726228305\") " Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.004235 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c522429-96d7-44c9-812c-f86726228305-kube-api-access-4p9fw" (OuterVolumeSpecName: "kube-api-access-4p9fw") pod "9c522429-96d7-44c9-812c-f86726228305" (UID: "9c522429-96d7-44c9-812c-f86726228305"). InnerVolumeSpecName "kube-api-access-4p9fw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.020396 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "9c522429-96d7-44c9-812c-f86726228305" (UID: "9c522429-96d7-44c9-812c-f86726228305"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.052207 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9c522429-96d7-44c9-812c-f86726228305" (UID: "9c522429-96d7-44c9-812c-f86726228305"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.055286 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-inventory" (OuterVolumeSpecName: "inventory") pod "9c522429-96d7-44c9-812c-f86726228305" (UID: "9c522429-96d7-44c9-812c-f86726228305"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.080709 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c522429-96d7-44c9-812c-f86726228305-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "9c522429-96d7-44c9-812c-f86726228305" (UID: "9c522429-96d7-44c9-812c-f86726228305"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.083793 4777 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9c522429-96d7-44c9-812c-f86726228305-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.083818 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.083827 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4p9fw\" (UniqueName: \"kubernetes.io/projected/9c522429-96d7-44c9-812c-f86726228305-kube-api-access-4p9fw\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.083839 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.083848 4777 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c522429-96d7-44c9-812c-f86726228305-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.307580 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" event={"ID":"9c522429-96d7-44c9-812c-f86726228305","Type":"ContainerDied","Data":"4caad6c443d78dd20f955704d44832dfc1788f048a3efd35342c27fe949f7f9a"} Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.308156 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4caad6c443d78dd20f955704d44832dfc1788f048a3efd35342c27fe949f7f9a" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.308307 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkd2c" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.414778 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2"] Nov 24 17:36:23 crc kubenswrapper[4777]: E1124 17:36:23.415333 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c522429-96d7-44c9-812c-f86726228305" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.415351 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c522429-96d7-44c9-812c-f86726228305" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 17:36:23 crc kubenswrapper[4777]: E1124 17:36:23.415371 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e693b21-413a-4ae2-bbfd-8b860359a74e" containerName="registry-server" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.415378 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e693b21-413a-4ae2-bbfd-8b860359a74e" containerName="registry-server" Nov 24 17:36:23 crc kubenswrapper[4777]: E1124 17:36:23.415399 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e693b21-413a-4ae2-bbfd-8b860359a74e" containerName="extract-content" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.415406 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e693b21-413a-4ae2-bbfd-8b860359a74e" containerName="extract-content" Nov 24 17:36:23 crc kubenswrapper[4777]: E1124 17:36:23.415419 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e693b21-413a-4ae2-bbfd-8b860359a74e" containerName="extract-utilities" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.415426 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e693b21-413a-4ae2-bbfd-8b860359a74e" containerName="extract-utilities" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.415628 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c522429-96d7-44c9-812c-f86726228305" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.415660 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e693b21-413a-4ae2-bbfd-8b860359a74e" containerName="registry-server" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.416376 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.418764 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.419057 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.419090 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.419148 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.419154 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.420857 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.431080 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2"] Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.597527 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.597569 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.597661 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.597694 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.597732 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.597785 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tblml\" (UniqueName: \"kubernetes.io/projected/97805347-c06c-409d-a299-cd4cfdf7e98e-kube-api-access-tblml\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.699099 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.699453 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.699562 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.699714 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tblml\" (UniqueName: \"kubernetes.io/projected/97805347-c06c-409d-a299-cd4cfdf7e98e-kube-api-access-tblml\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.699920 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.700046 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.704162 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.706839 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.707036 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.708585 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.709624 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.725144 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tblml\" (UniqueName: \"kubernetes.io/projected/97805347-c06c-409d-a299-cd4cfdf7e98e-kube-api-access-tblml\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:23 crc kubenswrapper[4777]: I1124 17:36:23.737124 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:36:24 crc kubenswrapper[4777]: I1124 17:36:24.332430 4777 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:36:24 crc kubenswrapper[4777]: I1124 17:36:24.337186 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2"] Nov 24 17:36:25 crc kubenswrapper[4777]: I1124 17:36:25.330218 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" event={"ID":"97805347-c06c-409d-a299-cd4cfdf7e98e","Type":"ContainerStarted","Data":"341adbc02f7cdf3d0b04adca4bb375151418f5c0ada490a8edda42e9cde4db97"} Nov 24 17:36:26 crc kubenswrapper[4777]: I1124 17:36:26.343794 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" event={"ID":"97805347-c06c-409d-a299-cd4cfdf7e98e","Type":"ContainerStarted","Data":"78fca53b6391671758e9fdc922877903d15ebd3dc1e5eaca1410781cdaa63fe0"} Nov 24 17:36:26 crc kubenswrapper[4777]: I1124 17:36:26.378261 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" podStartSLOduration=2.411652858 podStartE2EDuration="3.378236722s" podCreationTimestamp="2025-11-24 17:36:23 +0000 UTC" firstStartedPulling="2025-11-24 17:36:24.332115875 +0000 UTC m=+2142.491310924" lastFinishedPulling="2025-11-24 17:36:25.298699739 +0000 UTC m=+2143.457894788" observedRunningTime="2025-11-24 17:36:26.36246179 +0000 UTC m=+2144.521656869" watchObservedRunningTime="2025-11-24 17:36:26.378236722 +0000 UTC m=+2144.537431771" Nov 24 17:36:44 crc kubenswrapper[4777]: I1124 17:36:44.270057 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:36:44 crc kubenswrapper[4777]: I1124 17:36:44.270693 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:37:12 crc kubenswrapper[4777]: I1124 17:37:12.805449 4777 generic.go:334] "Generic (PLEG): container finished" podID="97805347-c06c-409d-a299-cd4cfdf7e98e" containerID="78fca53b6391671758e9fdc922877903d15ebd3dc1e5eaca1410781cdaa63fe0" exitCode=0 Nov 24 17:37:12 crc kubenswrapper[4777]: I1124 17:37:12.805533 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" event={"ID":"97805347-c06c-409d-a299-cd4cfdf7e98e","Type":"ContainerDied","Data":"78fca53b6391671758e9fdc922877903d15ebd3dc1e5eaca1410781cdaa63fe0"} Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.270929 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.271620 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.323568 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.399274 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"97805347-c06c-409d-a299-cd4cfdf7e98e\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.399551 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-neutron-metadata-combined-ca-bundle\") pod \"97805347-c06c-409d-a299-cd4cfdf7e98e\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.399641 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-inventory\") pod \"97805347-c06c-409d-a299-cd4cfdf7e98e\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.399797 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-nova-metadata-neutron-config-0\") pod \"97805347-c06c-409d-a299-cd4cfdf7e98e\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.400093 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tblml\" (UniqueName: \"kubernetes.io/projected/97805347-c06c-409d-a299-cd4cfdf7e98e-kube-api-access-tblml\") pod \"97805347-c06c-409d-a299-cd4cfdf7e98e\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.400496 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-ssh-key\") pod \"97805347-c06c-409d-a299-cd4cfdf7e98e\" (UID: \"97805347-c06c-409d-a299-cd4cfdf7e98e\") " Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.417309 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "97805347-c06c-409d-a299-cd4cfdf7e98e" (UID: "97805347-c06c-409d-a299-cd4cfdf7e98e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.417333 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97805347-c06c-409d-a299-cd4cfdf7e98e-kube-api-access-tblml" (OuterVolumeSpecName: "kube-api-access-tblml") pod "97805347-c06c-409d-a299-cd4cfdf7e98e" (UID: "97805347-c06c-409d-a299-cd4cfdf7e98e"). InnerVolumeSpecName "kube-api-access-tblml". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.428651 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "97805347-c06c-409d-a299-cd4cfdf7e98e" (UID: "97805347-c06c-409d-a299-cd4cfdf7e98e"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.431254 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "97805347-c06c-409d-a299-cd4cfdf7e98e" (UID: "97805347-c06c-409d-a299-cd4cfdf7e98e"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.441469 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "97805347-c06c-409d-a299-cd4cfdf7e98e" (UID: "97805347-c06c-409d-a299-cd4cfdf7e98e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.451199 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-inventory" (OuterVolumeSpecName: "inventory") pod "97805347-c06c-409d-a299-cd4cfdf7e98e" (UID: "97805347-c06c-409d-a299-cd4cfdf7e98e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.504335 4777 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.504588 4777 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.504709 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.504792 4777 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.504881 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tblml\" (UniqueName: \"kubernetes.io/projected/97805347-c06c-409d-a299-cd4cfdf7e98e-kube-api-access-tblml\") on node \"crc\" DevicePath \"\"" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.505051 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97805347-c06c-409d-a299-cd4cfdf7e98e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.827212 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" event={"ID":"97805347-c06c-409d-a299-cd4cfdf7e98e","Type":"ContainerDied","Data":"341adbc02f7cdf3d0b04adca4bb375151418f5c0ada490a8edda42e9cde4db97"} Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.827266 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="341adbc02f7cdf3d0b04adca4bb375151418f5c0ada490a8edda42e9cde4db97" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.827294 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.925725 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792"] Nov 24 17:37:14 crc kubenswrapper[4777]: E1124 17:37:14.926239 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97805347-c06c-409d-a299-cd4cfdf7e98e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.926264 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="97805347-c06c-409d-a299-cd4cfdf7e98e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.926576 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="97805347-c06c-409d-a299-cd4cfdf7e98e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.927495 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.929411 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.929808 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.930125 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.930350 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.930547 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:37:14 crc kubenswrapper[4777]: I1124 17:37:14.965615 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792"] Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.015308 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5frmt\" (UniqueName: \"kubernetes.io/projected/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-kube-api-access-5frmt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-md792\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.015385 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-md792\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.015466 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-md792\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.015525 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-md792\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.015609 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-md792\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.117466 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-md792\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.117912 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5frmt\" (UniqueName: \"kubernetes.io/projected/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-kube-api-access-5frmt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-md792\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.118001 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-md792\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.118050 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-md792\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.118111 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-md792\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.128821 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-md792\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.129015 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-md792\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.132622 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-md792\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.133822 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-md792\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.146463 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5frmt\" (UniqueName: \"kubernetes.io/projected/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-kube-api-access-5frmt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-md792\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.292780 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:37:15 crc kubenswrapper[4777]: I1124 17:37:15.831224 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792"] Nov 24 17:37:16 crc kubenswrapper[4777]: I1124 17:37:16.874110 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" event={"ID":"fd79eb9e-a6a2-4932-8a81-27a115ccaef6","Type":"ContainerStarted","Data":"3bbd03f652ec8d49be046d36164bbd5137dea5b1cac033f187a122e0c0b860bd"} Nov 24 17:37:17 crc kubenswrapper[4777]: I1124 17:37:17.886224 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" event={"ID":"fd79eb9e-a6a2-4932-8a81-27a115ccaef6","Type":"ContainerStarted","Data":"279d7cb8472caa512ff4ab3b53a03f68ddea92375f935bc3c1caa3956c0fd44d"} Nov 24 17:37:17 crc kubenswrapper[4777]: I1124 17:37:17.909664 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" podStartSLOduration=2.7567843549999997 podStartE2EDuration="3.909640156s" podCreationTimestamp="2025-11-24 17:37:14 +0000 UTC" firstStartedPulling="2025-11-24 17:37:15.842769276 +0000 UTC m=+2194.001964325" lastFinishedPulling="2025-11-24 17:37:16.995625077 +0000 UTC m=+2195.154820126" observedRunningTime="2025-11-24 17:37:17.899626425 +0000 UTC m=+2196.058821474" watchObservedRunningTime="2025-11-24 17:37:17.909640156 +0000 UTC m=+2196.068835205" Nov 24 17:37:44 crc kubenswrapper[4777]: I1124 17:37:44.270887 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:37:44 crc kubenswrapper[4777]: I1124 17:37:44.271445 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:37:44 crc kubenswrapper[4777]: I1124 17:37:44.271486 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:37:44 crc kubenswrapper[4777]: I1124 17:37:44.272354 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:37:44 crc kubenswrapper[4777]: I1124 17:37:44.272427 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" gracePeriod=600 Nov 24 17:37:44 crc kubenswrapper[4777]: E1124 17:37:44.399398 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:37:45 crc kubenswrapper[4777]: I1124 17:37:45.225926 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" exitCode=0 Nov 24 17:37:45 crc kubenswrapper[4777]: I1124 17:37:45.226529 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d"} Nov 24 17:37:45 crc kubenswrapper[4777]: I1124 17:37:45.226597 4777 scope.go:117] "RemoveContainer" containerID="a52a61b464393ded65c18d60855e7e57f0df7e2cf82fdbb47caf6735df256f42" Nov 24 17:37:45 crc kubenswrapper[4777]: I1124 17:37:45.227426 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:37:45 crc kubenswrapper[4777]: E1124 17:37:45.227843 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:37:59 crc kubenswrapper[4777]: I1124 17:37:59.245846 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:37:59 crc kubenswrapper[4777]: E1124 17:37:59.246935 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:38:12 crc kubenswrapper[4777]: I1124 17:38:12.047586 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q4m7q"] Nov 24 17:38:12 crc kubenswrapper[4777]: I1124 17:38:12.051311 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:12 crc kubenswrapper[4777]: I1124 17:38:12.062896 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q4m7q"] Nov 24 17:38:12 crc kubenswrapper[4777]: I1124 17:38:12.212290 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf66a177-86f6-4b43-ba68-359d231c9535-utilities\") pod \"redhat-operators-q4m7q\" (UID: \"cf66a177-86f6-4b43-ba68-359d231c9535\") " pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:12 crc kubenswrapper[4777]: I1124 17:38:12.212386 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr2k7\" (UniqueName: \"kubernetes.io/projected/cf66a177-86f6-4b43-ba68-359d231c9535-kube-api-access-cr2k7\") pod \"redhat-operators-q4m7q\" (UID: \"cf66a177-86f6-4b43-ba68-359d231c9535\") " pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:12 crc kubenswrapper[4777]: I1124 17:38:12.212443 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf66a177-86f6-4b43-ba68-359d231c9535-catalog-content\") pod \"redhat-operators-q4m7q\" (UID: \"cf66a177-86f6-4b43-ba68-359d231c9535\") " pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:12 crc kubenswrapper[4777]: I1124 17:38:12.245076 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:38:12 crc kubenswrapper[4777]: E1124 17:38:12.245492 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:38:12 crc kubenswrapper[4777]: I1124 17:38:12.313978 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr2k7\" (UniqueName: \"kubernetes.io/projected/cf66a177-86f6-4b43-ba68-359d231c9535-kube-api-access-cr2k7\") pod \"redhat-operators-q4m7q\" (UID: \"cf66a177-86f6-4b43-ba68-359d231c9535\") " pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:12 crc kubenswrapper[4777]: I1124 17:38:12.314092 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf66a177-86f6-4b43-ba68-359d231c9535-catalog-content\") pod \"redhat-operators-q4m7q\" (UID: \"cf66a177-86f6-4b43-ba68-359d231c9535\") " pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:12 crc kubenswrapper[4777]: I1124 17:38:12.314192 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf66a177-86f6-4b43-ba68-359d231c9535-utilities\") pod \"redhat-operators-q4m7q\" (UID: \"cf66a177-86f6-4b43-ba68-359d231c9535\") " pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:12 crc kubenswrapper[4777]: I1124 17:38:12.314723 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf66a177-86f6-4b43-ba68-359d231c9535-utilities\") pod \"redhat-operators-q4m7q\" (UID: \"cf66a177-86f6-4b43-ba68-359d231c9535\") " pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:12 crc kubenswrapper[4777]: I1124 17:38:12.314784 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf66a177-86f6-4b43-ba68-359d231c9535-catalog-content\") pod \"redhat-operators-q4m7q\" (UID: \"cf66a177-86f6-4b43-ba68-359d231c9535\") " pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:12 crc kubenswrapper[4777]: I1124 17:38:12.334471 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr2k7\" (UniqueName: \"kubernetes.io/projected/cf66a177-86f6-4b43-ba68-359d231c9535-kube-api-access-cr2k7\") pod \"redhat-operators-q4m7q\" (UID: \"cf66a177-86f6-4b43-ba68-359d231c9535\") " pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:12 crc kubenswrapper[4777]: I1124 17:38:12.409815 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:12 crc kubenswrapper[4777]: I1124 17:38:12.865774 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q4m7q"] Nov 24 17:38:13 crc kubenswrapper[4777]: I1124 17:38:13.553361 4777 generic.go:334] "Generic (PLEG): container finished" podID="cf66a177-86f6-4b43-ba68-359d231c9535" containerID="556a8bf8d191269b4b9fb078d007c593aceb7b3b2fb5c1bda393acafe2f51fed" exitCode=0 Nov 24 17:38:13 crc kubenswrapper[4777]: I1124 17:38:13.553468 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q4m7q" event={"ID":"cf66a177-86f6-4b43-ba68-359d231c9535","Type":"ContainerDied","Data":"556a8bf8d191269b4b9fb078d007c593aceb7b3b2fb5c1bda393acafe2f51fed"} Nov 24 17:38:13 crc kubenswrapper[4777]: I1124 17:38:13.553691 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q4m7q" event={"ID":"cf66a177-86f6-4b43-ba68-359d231c9535","Type":"ContainerStarted","Data":"56c21d172c33e3eda11c4bf797a2dcfbb88a8aa65823c9706eb5c97a82ad8296"} Nov 24 17:38:15 crc kubenswrapper[4777]: I1124 17:38:15.579297 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q4m7q" event={"ID":"cf66a177-86f6-4b43-ba68-359d231c9535","Type":"ContainerStarted","Data":"20727da55141c9a90807acfd551ae2010d4da19389a63c074ba72e2882abd258"} Nov 24 17:38:19 crc kubenswrapper[4777]: I1124 17:38:19.621772 4777 generic.go:334] "Generic (PLEG): container finished" podID="cf66a177-86f6-4b43-ba68-359d231c9535" containerID="20727da55141c9a90807acfd551ae2010d4da19389a63c074ba72e2882abd258" exitCode=0 Nov 24 17:38:19 crc kubenswrapper[4777]: I1124 17:38:19.621843 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q4m7q" event={"ID":"cf66a177-86f6-4b43-ba68-359d231c9535","Type":"ContainerDied","Data":"20727da55141c9a90807acfd551ae2010d4da19389a63c074ba72e2882abd258"} Nov 24 17:38:20 crc kubenswrapper[4777]: I1124 17:38:20.635198 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q4m7q" event={"ID":"cf66a177-86f6-4b43-ba68-359d231c9535","Type":"ContainerStarted","Data":"3536c0a152613fd4935a57ee1ff7d28f00d5bfa1247b647d70fbd52c8803966d"} Nov 24 17:38:20 crc kubenswrapper[4777]: I1124 17:38:20.661007 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q4m7q" podStartSLOduration=2.172320328 podStartE2EDuration="8.660987312s" podCreationTimestamp="2025-11-24 17:38:12 +0000 UTC" firstStartedPulling="2025-11-24 17:38:13.554719128 +0000 UTC m=+2251.713914197" lastFinishedPulling="2025-11-24 17:38:20.043386112 +0000 UTC m=+2258.202581181" observedRunningTime="2025-11-24 17:38:20.651594109 +0000 UTC m=+2258.810789158" watchObservedRunningTime="2025-11-24 17:38:20.660987312 +0000 UTC m=+2258.820182361" Nov 24 17:38:22 crc kubenswrapper[4777]: I1124 17:38:22.410898 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:22 crc kubenswrapper[4777]: I1124 17:38:22.411325 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:23 crc kubenswrapper[4777]: I1124 17:38:23.258323 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:38:23 crc kubenswrapper[4777]: E1124 17:38:23.261789 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:38:23 crc kubenswrapper[4777]: I1124 17:38:23.463005 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q4m7q" podUID="cf66a177-86f6-4b43-ba68-359d231c9535" containerName="registry-server" probeResult="failure" output=< Nov 24 17:38:23 crc kubenswrapper[4777]: timeout: failed to connect service ":50051" within 1s Nov 24 17:38:23 crc kubenswrapper[4777]: > Nov 24 17:38:32 crc kubenswrapper[4777]: I1124 17:38:32.468780 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:32 crc kubenswrapper[4777]: I1124 17:38:32.537804 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:32 crc kubenswrapper[4777]: I1124 17:38:32.708591 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q4m7q"] Nov 24 17:38:33 crc kubenswrapper[4777]: I1124 17:38:33.774863 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q4m7q" podUID="cf66a177-86f6-4b43-ba68-359d231c9535" containerName="registry-server" containerID="cri-o://3536c0a152613fd4935a57ee1ff7d28f00d5bfa1247b647d70fbd52c8803966d" gracePeriod=2 Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.333247 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.499684 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf66a177-86f6-4b43-ba68-359d231c9535-utilities\") pod \"cf66a177-86f6-4b43-ba68-359d231c9535\" (UID: \"cf66a177-86f6-4b43-ba68-359d231c9535\") " Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.499798 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cr2k7\" (UniqueName: \"kubernetes.io/projected/cf66a177-86f6-4b43-ba68-359d231c9535-kube-api-access-cr2k7\") pod \"cf66a177-86f6-4b43-ba68-359d231c9535\" (UID: \"cf66a177-86f6-4b43-ba68-359d231c9535\") " Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.499914 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf66a177-86f6-4b43-ba68-359d231c9535-catalog-content\") pod \"cf66a177-86f6-4b43-ba68-359d231c9535\" (UID: \"cf66a177-86f6-4b43-ba68-359d231c9535\") " Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.500664 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf66a177-86f6-4b43-ba68-359d231c9535-utilities" (OuterVolumeSpecName: "utilities") pod "cf66a177-86f6-4b43-ba68-359d231c9535" (UID: "cf66a177-86f6-4b43-ba68-359d231c9535"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.507842 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf66a177-86f6-4b43-ba68-359d231c9535-kube-api-access-cr2k7" (OuterVolumeSpecName: "kube-api-access-cr2k7") pod "cf66a177-86f6-4b43-ba68-359d231c9535" (UID: "cf66a177-86f6-4b43-ba68-359d231c9535"). InnerVolumeSpecName "kube-api-access-cr2k7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.594951 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf66a177-86f6-4b43-ba68-359d231c9535-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf66a177-86f6-4b43-ba68-359d231c9535" (UID: "cf66a177-86f6-4b43-ba68-359d231c9535"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.602219 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf66a177-86f6-4b43-ba68-359d231c9535-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.602250 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cr2k7\" (UniqueName: \"kubernetes.io/projected/cf66a177-86f6-4b43-ba68-359d231c9535-kube-api-access-cr2k7\") on node \"crc\" DevicePath \"\"" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.602264 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf66a177-86f6-4b43-ba68-359d231c9535-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.786571 4777 generic.go:334] "Generic (PLEG): container finished" podID="cf66a177-86f6-4b43-ba68-359d231c9535" containerID="3536c0a152613fd4935a57ee1ff7d28f00d5bfa1247b647d70fbd52c8803966d" exitCode=0 Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.786607 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q4m7q" event={"ID":"cf66a177-86f6-4b43-ba68-359d231c9535","Type":"ContainerDied","Data":"3536c0a152613fd4935a57ee1ff7d28f00d5bfa1247b647d70fbd52c8803966d"} Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.786633 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q4m7q" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.786669 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q4m7q" event={"ID":"cf66a177-86f6-4b43-ba68-359d231c9535","Type":"ContainerDied","Data":"56c21d172c33e3eda11c4bf797a2dcfbb88a8aa65823c9706eb5c97a82ad8296"} Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.786694 4777 scope.go:117] "RemoveContainer" containerID="3536c0a152613fd4935a57ee1ff7d28f00d5bfa1247b647d70fbd52c8803966d" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.805781 4777 scope.go:117] "RemoveContainer" containerID="20727da55141c9a90807acfd551ae2010d4da19389a63c074ba72e2882abd258" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.822043 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q4m7q"] Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.832559 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q4m7q"] Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.851237 4777 scope.go:117] "RemoveContainer" containerID="556a8bf8d191269b4b9fb078d007c593aceb7b3b2fb5c1bda393acafe2f51fed" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.883145 4777 scope.go:117] "RemoveContainer" containerID="3536c0a152613fd4935a57ee1ff7d28f00d5bfa1247b647d70fbd52c8803966d" Nov 24 17:38:34 crc kubenswrapper[4777]: E1124 17:38:34.883532 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3536c0a152613fd4935a57ee1ff7d28f00d5bfa1247b647d70fbd52c8803966d\": container with ID starting with 3536c0a152613fd4935a57ee1ff7d28f00d5bfa1247b647d70fbd52c8803966d not found: ID does not exist" containerID="3536c0a152613fd4935a57ee1ff7d28f00d5bfa1247b647d70fbd52c8803966d" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.883569 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3536c0a152613fd4935a57ee1ff7d28f00d5bfa1247b647d70fbd52c8803966d"} err="failed to get container status \"3536c0a152613fd4935a57ee1ff7d28f00d5bfa1247b647d70fbd52c8803966d\": rpc error: code = NotFound desc = could not find container \"3536c0a152613fd4935a57ee1ff7d28f00d5bfa1247b647d70fbd52c8803966d\": container with ID starting with 3536c0a152613fd4935a57ee1ff7d28f00d5bfa1247b647d70fbd52c8803966d not found: ID does not exist" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.883594 4777 scope.go:117] "RemoveContainer" containerID="20727da55141c9a90807acfd551ae2010d4da19389a63c074ba72e2882abd258" Nov 24 17:38:34 crc kubenswrapper[4777]: E1124 17:38:34.884017 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20727da55141c9a90807acfd551ae2010d4da19389a63c074ba72e2882abd258\": container with ID starting with 20727da55141c9a90807acfd551ae2010d4da19389a63c074ba72e2882abd258 not found: ID does not exist" containerID="20727da55141c9a90807acfd551ae2010d4da19389a63c074ba72e2882abd258" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.884051 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20727da55141c9a90807acfd551ae2010d4da19389a63c074ba72e2882abd258"} err="failed to get container status \"20727da55141c9a90807acfd551ae2010d4da19389a63c074ba72e2882abd258\": rpc error: code = NotFound desc = could not find container \"20727da55141c9a90807acfd551ae2010d4da19389a63c074ba72e2882abd258\": container with ID starting with 20727da55141c9a90807acfd551ae2010d4da19389a63c074ba72e2882abd258 not found: ID does not exist" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.884100 4777 scope.go:117] "RemoveContainer" containerID="556a8bf8d191269b4b9fb078d007c593aceb7b3b2fb5c1bda393acafe2f51fed" Nov 24 17:38:34 crc kubenswrapper[4777]: E1124 17:38:34.884352 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"556a8bf8d191269b4b9fb078d007c593aceb7b3b2fb5c1bda393acafe2f51fed\": container with ID starting with 556a8bf8d191269b4b9fb078d007c593aceb7b3b2fb5c1bda393acafe2f51fed not found: ID does not exist" containerID="556a8bf8d191269b4b9fb078d007c593aceb7b3b2fb5c1bda393acafe2f51fed" Nov 24 17:38:34 crc kubenswrapper[4777]: I1124 17:38:34.884383 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"556a8bf8d191269b4b9fb078d007c593aceb7b3b2fb5c1bda393acafe2f51fed"} err="failed to get container status \"556a8bf8d191269b4b9fb078d007c593aceb7b3b2fb5c1bda393acafe2f51fed\": rpc error: code = NotFound desc = could not find container \"556a8bf8d191269b4b9fb078d007c593aceb7b3b2fb5c1bda393acafe2f51fed\": container with ID starting with 556a8bf8d191269b4b9fb078d007c593aceb7b3b2fb5c1bda393acafe2f51fed not found: ID does not exist" Nov 24 17:38:35 crc kubenswrapper[4777]: I1124 17:38:35.256916 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf66a177-86f6-4b43-ba68-359d231c9535" path="/var/lib/kubelet/pods/cf66a177-86f6-4b43-ba68-359d231c9535/volumes" Nov 24 17:38:37 crc kubenswrapper[4777]: I1124 17:38:37.245860 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:38:37 crc kubenswrapper[4777]: E1124 17:38:37.246678 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:38:48 crc kubenswrapper[4777]: I1124 17:38:48.246113 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:38:48 crc kubenswrapper[4777]: E1124 17:38:48.246854 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:39:00 crc kubenswrapper[4777]: I1124 17:39:00.245004 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:39:00 crc kubenswrapper[4777]: E1124 17:39:00.245763 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:39:14 crc kubenswrapper[4777]: I1124 17:39:14.245735 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:39:14 crc kubenswrapper[4777]: E1124 17:39:14.246686 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:39:27 crc kubenswrapper[4777]: I1124 17:39:27.246451 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:39:27 crc kubenswrapper[4777]: E1124 17:39:27.247796 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:39:39 crc kubenswrapper[4777]: I1124 17:39:39.246891 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:39:39 crc kubenswrapper[4777]: E1124 17:39:39.247764 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:39:52 crc kubenswrapper[4777]: I1124 17:39:52.245783 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:39:52 crc kubenswrapper[4777]: E1124 17:39:52.246859 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:40:03 crc kubenswrapper[4777]: I1124 17:40:03.251821 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:40:03 crc kubenswrapper[4777]: E1124 17:40:03.252677 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:40:16 crc kubenswrapper[4777]: I1124 17:40:16.245189 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:40:16 crc kubenswrapper[4777]: E1124 17:40:16.246189 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:40:28 crc kubenswrapper[4777]: I1124 17:40:28.244854 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:40:28 crc kubenswrapper[4777]: E1124 17:40:28.245644 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:40:39 crc kubenswrapper[4777]: I1124 17:40:39.245515 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:40:39 crc kubenswrapper[4777]: E1124 17:40:39.246328 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:40:54 crc kubenswrapper[4777]: I1124 17:40:54.245789 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:40:54 crc kubenswrapper[4777]: E1124 17:40:54.246540 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:41:09 crc kubenswrapper[4777]: I1124 17:41:09.245747 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:41:09 crc kubenswrapper[4777]: E1124 17:41:09.247025 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:41:24 crc kubenswrapper[4777]: I1124 17:41:24.246091 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:41:24 crc kubenswrapper[4777]: E1124 17:41:24.247373 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:41:35 crc kubenswrapper[4777]: I1124 17:41:35.791084 4777 generic.go:334] "Generic (PLEG): container finished" podID="fd79eb9e-a6a2-4932-8a81-27a115ccaef6" containerID="279d7cb8472caa512ff4ab3b53a03f68ddea92375f935bc3c1caa3956c0fd44d" exitCode=0 Nov 24 17:41:35 crc kubenswrapper[4777]: I1124 17:41:35.791211 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" event={"ID":"fd79eb9e-a6a2-4932-8a81-27a115ccaef6","Type":"ContainerDied","Data":"279d7cb8472caa512ff4ab3b53a03f68ddea92375f935bc3c1caa3956c0fd44d"} Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.340192 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.537621 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-inventory\") pod \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.538051 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-libvirt-secret-0\") pod \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.538124 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-ssh-key\") pod \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.538233 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-libvirt-combined-ca-bundle\") pod \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.538366 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5frmt\" (UniqueName: \"kubernetes.io/projected/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-kube-api-access-5frmt\") pod \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\" (UID: \"fd79eb9e-a6a2-4932-8a81-27a115ccaef6\") " Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.544705 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "fd79eb9e-a6a2-4932-8a81-27a115ccaef6" (UID: "fd79eb9e-a6a2-4932-8a81-27a115ccaef6"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.545031 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-kube-api-access-5frmt" (OuterVolumeSpecName: "kube-api-access-5frmt") pod "fd79eb9e-a6a2-4932-8a81-27a115ccaef6" (UID: "fd79eb9e-a6a2-4932-8a81-27a115ccaef6"). InnerVolumeSpecName "kube-api-access-5frmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.575087 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-inventory" (OuterVolumeSpecName: "inventory") pod "fd79eb9e-a6a2-4932-8a81-27a115ccaef6" (UID: "fd79eb9e-a6a2-4932-8a81-27a115ccaef6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.577294 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "fd79eb9e-a6a2-4932-8a81-27a115ccaef6" (UID: "fd79eb9e-a6a2-4932-8a81-27a115ccaef6"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.598113 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fd79eb9e-a6a2-4932-8a81-27a115ccaef6" (UID: "fd79eb9e-a6a2-4932-8a81-27a115ccaef6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.641019 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.641048 4777 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.641058 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.641069 4777 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.641078 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5frmt\" (UniqueName: \"kubernetes.io/projected/fd79eb9e-a6a2-4932-8a81-27a115ccaef6-kube-api-access-5frmt\") on node \"crc\" DevicePath \"\"" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.814793 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" event={"ID":"fd79eb9e-a6a2-4932-8a81-27a115ccaef6","Type":"ContainerDied","Data":"3bbd03f652ec8d49be046d36164bbd5137dea5b1cac033f187a122e0c0b860bd"} Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.814838 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bbd03f652ec8d49be046d36164bbd5137dea5b1cac033f187a122e0c0b860bd" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.814889 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-md792" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.910617 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5"] Nov 24 17:41:37 crc kubenswrapper[4777]: E1124 17:41:37.911617 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd79eb9e-a6a2-4932-8a81-27a115ccaef6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.911774 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd79eb9e-a6a2-4932-8a81-27a115ccaef6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 17:41:37 crc kubenswrapper[4777]: E1124 17:41:37.911923 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf66a177-86f6-4b43-ba68-359d231c9535" containerName="extract-content" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.912072 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf66a177-86f6-4b43-ba68-359d231c9535" containerName="extract-content" Nov 24 17:41:37 crc kubenswrapper[4777]: E1124 17:41:37.912198 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf66a177-86f6-4b43-ba68-359d231c9535" containerName="extract-utilities" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.912296 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf66a177-86f6-4b43-ba68-359d231c9535" containerName="extract-utilities" Nov 24 17:41:37 crc kubenswrapper[4777]: E1124 17:41:37.912432 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf66a177-86f6-4b43-ba68-359d231c9535" containerName="registry-server" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.912555 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf66a177-86f6-4b43-ba68-359d231c9535" containerName="registry-server" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.913070 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd79eb9e-a6a2-4932-8a81-27a115ccaef6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.913234 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf66a177-86f6-4b43-ba68-359d231c9535" containerName="registry-server" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.914788 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.917063 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.917545 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.919677 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.919851 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.919955 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.919940 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.920412 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.922821 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5"] Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.947129 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzjdh\" (UniqueName: \"kubernetes.io/projected/9651259d-da1a-4580-855a-94a9310fc952-kube-api-access-zzjdh\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.947422 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.947616 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.947732 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.947873 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.948012 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.948148 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.948383 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:37 crc kubenswrapper[4777]: I1124 17:41:37.948508 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9651259d-da1a-4580-855a-94a9310fc952-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.051040 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.051096 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.051144 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.051184 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.051224 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.051319 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.051348 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9651259d-da1a-4580-855a-94a9310fc952-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.051394 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzjdh\" (UniqueName: \"kubernetes.io/projected/9651259d-da1a-4580-855a-94a9310fc952-kube-api-access-zzjdh\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.051430 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.052958 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9651259d-da1a-4580-855a-94a9310fc952-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.054957 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.056440 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.057653 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.059327 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.062948 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.063333 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.067586 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.074312 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzjdh\" (UniqueName: \"kubernetes.io/projected/9651259d-da1a-4580-855a-94a9310fc952-kube-api-access-zzjdh\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w44w5\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.237947 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.246680 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:41:38 crc kubenswrapper[4777]: E1124 17:41:38.247101 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.866317 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5"] Nov 24 17:41:38 crc kubenswrapper[4777]: I1124 17:41:38.874018 4777 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:41:39 crc kubenswrapper[4777]: I1124 17:41:39.834541 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" event={"ID":"9651259d-da1a-4580-855a-94a9310fc952","Type":"ContainerStarted","Data":"0f500f670cf2023dce5ae10d72dcb21aa9663adf2a8bdaf6c16fdf2a3d9486ea"} Nov 24 17:41:39 crc kubenswrapper[4777]: I1124 17:41:39.834930 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" event={"ID":"9651259d-da1a-4580-855a-94a9310fc952","Type":"ContainerStarted","Data":"75f857f14624805e9a86d248a55fa49b493451fc52a0d45c2fe39e01d81f1529"} Nov 24 17:41:39 crc kubenswrapper[4777]: I1124 17:41:39.861340 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" podStartSLOduration=2.353040572 podStartE2EDuration="2.861324768s" podCreationTimestamp="2025-11-24 17:41:37 +0000 UTC" firstStartedPulling="2025-11-24 17:41:38.873565514 +0000 UTC m=+2457.032760563" lastFinishedPulling="2025-11-24 17:41:39.38184969 +0000 UTC m=+2457.541044759" observedRunningTime="2025-11-24 17:41:39.855327418 +0000 UTC m=+2458.014522477" watchObservedRunningTime="2025-11-24 17:41:39.861324768 +0000 UTC m=+2458.020519817" Nov 24 17:41:50 crc kubenswrapper[4777]: I1124 17:41:50.246511 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:41:50 crc kubenswrapper[4777]: E1124 17:41:50.249697 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:42:03 crc kubenswrapper[4777]: I1124 17:42:03.268632 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:42:03 crc kubenswrapper[4777]: E1124 17:42:03.269935 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:42:17 crc kubenswrapper[4777]: I1124 17:42:17.246034 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:42:17 crc kubenswrapper[4777]: E1124 17:42:17.248378 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:42:31 crc kubenswrapper[4777]: I1124 17:42:31.246468 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:42:31 crc kubenswrapper[4777]: E1124 17:42:31.247734 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:42:45 crc kubenswrapper[4777]: I1124 17:42:45.246474 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:42:45 crc kubenswrapper[4777]: I1124 17:42:45.612453 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"535d5108efbbfad99f1fa015e190845054474327b1f12cffe94f88b32d9cf532"} Nov 24 17:43:13 crc kubenswrapper[4777]: I1124 17:43:13.354608 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qfwx9"] Nov 24 17:43:13 crc kubenswrapper[4777]: I1124 17:43:13.358011 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:13 crc kubenswrapper[4777]: I1124 17:43:13.367030 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qfwx9"] Nov 24 17:43:13 crc kubenswrapper[4777]: I1124 17:43:13.473869 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d1f3579-f405-47c7-9728-f0a8508d88ac-utilities\") pod \"certified-operators-qfwx9\" (UID: \"3d1f3579-f405-47c7-9728-f0a8508d88ac\") " pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:13 crc kubenswrapper[4777]: I1124 17:43:13.474333 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z75q5\" (UniqueName: \"kubernetes.io/projected/3d1f3579-f405-47c7-9728-f0a8508d88ac-kube-api-access-z75q5\") pod \"certified-operators-qfwx9\" (UID: \"3d1f3579-f405-47c7-9728-f0a8508d88ac\") " pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:13 crc kubenswrapper[4777]: I1124 17:43:13.474360 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d1f3579-f405-47c7-9728-f0a8508d88ac-catalog-content\") pod \"certified-operators-qfwx9\" (UID: \"3d1f3579-f405-47c7-9728-f0a8508d88ac\") " pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:13 crc kubenswrapper[4777]: I1124 17:43:13.575826 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z75q5\" (UniqueName: \"kubernetes.io/projected/3d1f3579-f405-47c7-9728-f0a8508d88ac-kube-api-access-z75q5\") pod \"certified-operators-qfwx9\" (UID: \"3d1f3579-f405-47c7-9728-f0a8508d88ac\") " pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:13 crc kubenswrapper[4777]: I1124 17:43:13.575885 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d1f3579-f405-47c7-9728-f0a8508d88ac-catalog-content\") pod \"certified-operators-qfwx9\" (UID: \"3d1f3579-f405-47c7-9728-f0a8508d88ac\") " pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:13 crc kubenswrapper[4777]: I1124 17:43:13.576055 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d1f3579-f405-47c7-9728-f0a8508d88ac-utilities\") pod \"certified-operators-qfwx9\" (UID: \"3d1f3579-f405-47c7-9728-f0a8508d88ac\") " pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:13 crc kubenswrapper[4777]: I1124 17:43:13.576446 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d1f3579-f405-47c7-9728-f0a8508d88ac-catalog-content\") pod \"certified-operators-qfwx9\" (UID: \"3d1f3579-f405-47c7-9728-f0a8508d88ac\") " pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:13 crc kubenswrapper[4777]: I1124 17:43:13.576499 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d1f3579-f405-47c7-9728-f0a8508d88ac-utilities\") pod \"certified-operators-qfwx9\" (UID: \"3d1f3579-f405-47c7-9728-f0a8508d88ac\") " pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:13 crc kubenswrapper[4777]: I1124 17:43:13.600186 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z75q5\" (UniqueName: \"kubernetes.io/projected/3d1f3579-f405-47c7-9728-f0a8508d88ac-kube-api-access-z75q5\") pod \"certified-operators-qfwx9\" (UID: \"3d1f3579-f405-47c7-9728-f0a8508d88ac\") " pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:13 crc kubenswrapper[4777]: I1124 17:43:13.685052 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:14 crc kubenswrapper[4777]: W1124 17:43:14.239981 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d1f3579_f405_47c7_9728_f0a8508d88ac.slice/crio-69fc339964f98e4074a1251fda254bd9d04f6f0d354fb76f3ba42520c11ecf2c WatchSource:0}: Error finding container 69fc339964f98e4074a1251fda254bd9d04f6f0d354fb76f3ba42520c11ecf2c: Status 404 returned error can't find the container with id 69fc339964f98e4074a1251fda254bd9d04f6f0d354fb76f3ba42520c11ecf2c Nov 24 17:43:14 crc kubenswrapper[4777]: I1124 17:43:14.245604 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qfwx9"] Nov 24 17:43:14 crc kubenswrapper[4777]: I1124 17:43:14.949343 4777 generic.go:334] "Generic (PLEG): container finished" podID="3d1f3579-f405-47c7-9728-f0a8508d88ac" containerID="3823a963b656c48fe9fc408d3088be301d07c06bb9d28876815cab8c5b5a272d" exitCode=0 Nov 24 17:43:14 crc kubenswrapper[4777]: I1124 17:43:14.949384 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qfwx9" event={"ID":"3d1f3579-f405-47c7-9728-f0a8508d88ac","Type":"ContainerDied","Data":"3823a963b656c48fe9fc408d3088be301d07c06bb9d28876815cab8c5b5a272d"} Nov 24 17:43:14 crc kubenswrapper[4777]: I1124 17:43:14.949408 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qfwx9" event={"ID":"3d1f3579-f405-47c7-9728-f0a8508d88ac","Type":"ContainerStarted","Data":"69fc339964f98e4074a1251fda254bd9d04f6f0d354fb76f3ba42520c11ecf2c"} Nov 24 17:43:16 crc kubenswrapper[4777]: I1124 17:43:16.979049 4777 generic.go:334] "Generic (PLEG): container finished" podID="3d1f3579-f405-47c7-9728-f0a8508d88ac" containerID="60181908a6e93be5c30901359e25775b9534bfb674e2a15ef57f66516ef7027a" exitCode=0 Nov 24 17:43:16 crc kubenswrapper[4777]: I1124 17:43:16.979184 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qfwx9" event={"ID":"3d1f3579-f405-47c7-9728-f0a8508d88ac","Type":"ContainerDied","Data":"60181908a6e93be5c30901359e25775b9534bfb674e2a15ef57f66516ef7027a"} Nov 24 17:43:17 crc kubenswrapper[4777]: I1124 17:43:17.999090 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qfwx9" event={"ID":"3d1f3579-f405-47c7-9728-f0a8508d88ac","Type":"ContainerStarted","Data":"a3446e781975572a1079b9b634d20d9717f390dda091e6dbe2e40776730a9b76"} Nov 24 17:43:23 crc kubenswrapper[4777]: I1124 17:43:23.685272 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:23 crc kubenswrapper[4777]: I1124 17:43:23.685974 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:23 crc kubenswrapper[4777]: I1124 17:43:23.765970 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:23 crc kubenswrapper[4777]: I1124 17:43:23.800554 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qfwx9" podStartSLOduration=8.277244787 podStartE2EDuration="10.800530242s" podCreationTimestamp="2025-11-24 17:43:13 +0000 UTC" firstStartedPulling="2025-11-24 17:43:14.953320581 +0000 UTC m=+2553.112515630" lastFinishedPulling="2025-11-24 17:43:17.476606026 +0000 UTC m=+2555.635801085" observedRunningTime="2025-11-24 17:43:18.016320193 +0000 UTC m=+2556.175515242" watchObservedRunningTime="2025-11-24 17:43:23.800530242 +0000 UTC m=+2561.959725311" Nov 24 17:43:24 crc kubenswrapper[4777]: I1124 17:43:24.127572 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:24 crc kubenswrapper[4777]: I1124 17:43:24.193618 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qfwx9"] Nov 24 17:43:26 crc kubenswrapper[4777]: I1124 17:43:26.092391 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qfwx9" podUID="3d1f3579-f405-47c7-9728-f0a8508d88ac" containerName="registry-server" containerID="cri-o://a3446e781975572a1079b9b634d20d9717f390dda091e6dbe2e40776730a9b76" gracePeriod=2 Nov 24 17:43:26 crc kubenswrapper[4777]: I1124 17:43:26.613409 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:26 crc kubenswrapper[4777]: I1124 17:43:26.766459 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d1f3579-f405-47c7-9728-f0a8508d88ac-utilities\") pod \"3d1f3579-f405-47c7-9728-f0a8508d88ac\" (UID: \"3d1f3579-f405-47c7-9728-f0a8508d88ac\") " Nov 24 17:43:26 crc kubenswrapper[4777]: I1124 17:43:26.766579 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z75q5\" (UniqueName: \"kubernetes.io/projected/3d1f3579-f405-47c7-9728-f0a8508d88ac-kube-api-access-z75q5\") pod \"3d1f3579-f405-47c7-9728-f0a8508d88ac\" (UID: \"3d1f3579-f405-47c7-9728-f0a8508d88ac\") " Nov 24 17:43:26 crc kubenswrapper[4777]: I1124 17:43:26.766866 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d1f3579-f405-47c7-9728-f0a8508d88ac-catalog-content\") pod \"3d1f3579-f405-47c7-9728-f0a8508d88ac\" (UID: \"3d1f3579-f405-47c7-9728-f0a8508d88ac\") " Nov 24 17:43:26 crc kubenswrapper[4777]: I1124 17:43:26.767811 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d1f3579-f405-47c7-9728-f0a8508d88ac-utilities" (OuterVolumeSpecName: "utilities") pod "3d1f3579-f405-47c7-9728-f0a8508d88ac" (UID: "3d1f3579-f405-47c7-9728-f0a8508d88ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:43:26 crc kubenswrapper[4777]: I1124 17:43:26.773130 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d1f3579-f405-47c7-9728-f0a8508d88ac-kube-api-access-z75q5" (OuterVolumeSpecName: "kube-api-access-z75q5") pod "3d1f3579-f405-47c7-9728-f0a8508d88ac" (UID: "3d1f3579-f405-47c7-9728-f0a8508d88ac"). InnerVolumeSpecName "kube-api-access-z75q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:43:26 crc kubenswrapper[4777]: I1124 17:43:26.870193 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d1f3579-f405-47c7-9728-f0a8508d88ac-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:26 crc kubenswrapper[4777]: I1124 17:43:26.870237 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z75q5\" (UniqueName: \"kubernetes.io/projected/3d1f3579-f405-47c7-9728-f0a8508d88ac-kube-api-access-z75q5\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.107160 4777 generic.go:334] "Generic (PLEG): container finished" podID="3d1f3579-f405-47c7-9728-f0a8508d88ac" containerID="a3446e781975572a1079b9b634d20d9717f390dda091e6dbe2e40776730a9b76" exitCode=0 Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.107225 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qfwx9" event={"ID":"3d1f3579-f405-47c7-9728-f0a8508d88ac","Type":"ContainerDied","Data":"a3446e781975572a1079b9b634d20d9717f390dda091e6dbe2e40776730a9b76"} Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.107264 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qfwx9" event={"ID":"3d1f3579-f405-47c7-9728-f0a8508d88ac","Type":"ContainerDied","Data":"69fc339964f98e4074a1251fda254bd9d04f6f0d354fb76f3ba42520c11ecf2c"} Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.107293 4777 scope.go:117] "RemoveContainer" containerID="a3446e781975572a1079b9b634d20d9717f390dda091e6dbe2e40776730a9b76" Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.107475 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qfwx9" Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.143502 4777 scope.go:117] "RemoveContainer" containerID="60181908a6e93be5c30901359e25775b9534bfb674e2a15ef57f66516ef7027a" Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.169223 4777 scope.go:117] "RemoveContainer" containerID="3823a963b656c48fe9fc408d3088be301d07c06bb9d28876815cab8c5b5a272d" Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.239174 4777 scope.go:117] "RemoveContainer" containerID="a3446e781975572a1079b9b634d20d9717f390dda091e6dbe2e40776730a9b76" Nov 24 17:43:27 crc kubenswrapper[4777]: E1124 17:43:27.240239 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3446e781975572a1079b9b634d20d9717f390dda091e6dbe2e40776730a9b76\": container with ID starting with a3446e781975572a1079b9b634d20d9717f390dda091e6dbe2e40776730a9b76 not found: ID does not exist" containerID="a3446e781975572a1079b9b634d20d9717f390dda091e6dbe2e40776730a9b76" Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.240305 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3446e781975572a1079b9b634d20d9717f390dda091e6dbe2e40776730a9b76"} err="failed to get container status \"a3446e781975572a1079b9b634d20d9717f390dda091e6dbe2e40776730a9b76\": rpc error: code = NotFound desc = could not find container \"a3446e781975572a1079b9b634d20d9717f390dda091e6dbe2e40776730a9b76\": container with ID starting with a3446e781975572a1079b9b634d20d9717f390dda091e6dbe2e40776730a9b76 not found: ID does not exist" Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.240346 4777 scope.go:117] "RemoveContainer" containerID="60181908a6e93be5c30901359e25775b9534bfb674e2a15ef57f66516ef7027a" Nov 24 17:43:27 crc kubenswrapper[4777]: E1124 17:43:27.240933 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60181908a6e93be5c30901359e25775b9534bfb674e2a15ef57f66516ef7027a\": container with ID starting with 60181908a6e93be5c30901359e25775b9534bfb674e2a15ef57f66516ef7027a not found: ID does not exist" containerID="60181908a6e93be5c30901359e25775b9534bfb674e2a15ef57f66516ef7027a" Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.241068 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60181908a6e93be5c30901359e25775b9534bfb674e2a15ef57f66516ef7027a"} err="failed to get container status \"60181908a6e93be5c30901359e25775b9534bfb674e2a15ef57f66516ef7027a\": rpc error: code = NotFound desc = could not find container \"60181908a6e93be5c30901359e25775b9534bfb674e2a15ef57f66516ef7027a\": container with ID starting with 60181908a6e93be5c30901359e25775b9534bfb674e2a15ef57f66516ef7027a not found: ID does not exist" Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.241096 4777 scope.go:117] "RemoveContainer" containerID="3823a963b656c48fe9fc408d3088be301d07c06bb9d28876815cab8c5b5a272d" Nov 24 17:43:27 crc kubenswrapper[4777]: E1124 17:43:27.241616 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3823a963b656c48fe9fc408d3088be301d07c06bb9d28876815cab8c5b5a272d\": container with ID starting with 3823a963b656c48fe9fc408d3088be301d07c06bb9d28876815cab8c5b5a272d not found: ID does not exist" containerID="3823a963b656c48fe9fc408d3088be301d07c06bb9d28876815cab8c5b5a272d" Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.241781 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3823a963b656c48fe9fc408d3088be301d07c06bb9d28876815cab8c5b5a272d"} err="failed to get container status \"3823a963b656c48fe9fc408d3088be301d07c06bb9d28876815cab8c5b5a272d\": rpc error: code = NotFound desc = could not find container \"3823a963b656c48fe9fc408d3088be301d07c06bb9d28876815cab8c5b5a272d\": container with ID starting with 3823a963b656c48fe9fc408d3088be301d07c06bb9d28876815cab8c5b5a272d not found: ID does not exist" Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.367235 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d1f3579-f405-47c7-9728-f0a8508d88ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d1f3579-f405-47c7-9728-f0a8508d88ac" (UID: "3d1f3579-f405-47c7-9728-f0a8508d88ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.384559 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d1f3579-f405-47c7-9728-f0a8508d88ac-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.453307 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qfwx9"] Nov 24 17:43:27 crc kubenswrapper[4777]: I1124 17:43:27.470484 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qfwx9"] Nov 24 17:43:29 crc kubenswrapper[4777]: I1124 17:43:29.258286 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d1f3579-f405-47c7-9728-f0a8508d88ac" path="/var/lib/kubelet/pods/3d1f3579-f405-47c7-9728-f0a8508d88ac/volumes" Nov 24 17:44:15 crc kubenswrapper[4777]: I1124 17:44:15.694995 4777 generic.go:334] "Generic (PLEG): container finished" podID="9651259d-da1a-4580-855a-94a9310fc952" containerID="0f500f670cf2023dce5ae10d72dcb21aa9663adf2a8bdaf6c16fdf2a3d9486ea" exitCode=0 Nov 24 17:44:15 crc kubenswrapper[4777]: I1124 17:44:15.695059 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" event={"ID":"9651259d-da1a-4580-855a-94a9310fc952","Type":"ContainerDied","Data":"0f500f670cf2023dce5ae10d72dcb21aa9663adf2a8bdaf6c16fdf2a3d9486ea"} Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.218685 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.409760 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-migration-ssh-key-1\") pod \"9651259d-da1a-4580-855a-94a9310fc952\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.409871 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-cell1-compute-config-1\") pod \"9651259d-da1a-4580-855a-94a9310fc952\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.409898 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-ssh-key\") pod \"9651259d-da1a-4580-855a-94a9310fc952\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.409995 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-cell1-compute-config-0\") pod \"9651259d-da1a-4580-855a-94a9310fc952\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.410035 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-migration-ssh-key-0\") pod \"9651259d-da1a-4580-855a-94a9310fc952\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.410063 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzjdh\" (UniqueName: \"kubernetes.io/projected/9651259d-da1a-4580-855a-94a9310fc952-kube-api-access-zzjdh\") pod \"9651259d-da1a-4580-855a-94a9310fc952\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.410165 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-inventory\") pod \"9651259d-da1a-4580-855a-94a9310fc952\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.410305 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-combined-ca-bundle\") pod \"9651259d-da1a-4580-855a-94a9310fc952\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.410351 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9651259d-da1a-4580-855a-94a9310fc952-nova-extra-config-0\") pod \"9651259d-da1a-4580-855a-94a9310fc952\" (UID: \"9651259d-da1a-4580-855a-94a9310fc952\") " Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.422922 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9651259d-da1a-4580-855a-94a9310fc952-kube-api-access-zzjdh" (OuterVolumeSpecName: "kube-api-access-zzjdh") pod "9651259d-da1a-4580-855a-94a9310fc952" (UID: "9651259d-da1a-4580-855a-94a9310fc952"). InnerVolumeSpecName "kube-api-access-zzjdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.432907 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "9651259d-da1a-4580-855a-94a9310fc952" (UID: "9651259d-da1a-4580-855a-94a9310fc952"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.451189 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "9651259d-da1a-4580-855a-94a9310fc952" (UID: "9651259d-da1a-4580-855a-94a9310fc952"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.455033 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9651259d-da1a-4580-855a-94a9310fc952-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "9651259d-da1a-4580-855a-94a9310fc952" (UID: "9651259d-da1a-4580-855a-94a9310fc952"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.465176 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "9651259d-da1a-4580-855a-94a9310fc952" (UID: "9651259d-da1a-4580-855a-94a9310fc952"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.471674 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9651259d-da1a-4580-855a-94a9310fc952" (UID: "9651259d-da1a-4580-855a-94a9310fc952"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.477191 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "9651259d-da1a-4580-855a-94a9310fc952" (UID: "9651259d-da1a-4580-855a-94a9310fc952"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.478593 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "9651259d-da1a-4580-855a-94a9310fc952" (UID: "9651259d-da1a-4580-855a-94a9310fc952"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.494390 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-inventory" (OuterVolumeSpecName: "inventory") pod "9651259d-da1a-4580-855a-94a9310fc952" (UID: "9651259d-da1a-4580-855a-94a9310fc952"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.512397 4777 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.512434 4777 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9651259d-da1a-4580-855a-94a9310fc952-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.512445 4777 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.512458 4777 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.512470 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.512483 4777 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.512493 4777 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.512505 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzjdh\" (UniqueName: \"kubernetes.io/projected/9651259d-da1a-4580-855a-94a9310fc952-kube-api-access-zzjdh\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.512516 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9651259d-da1a-4580-855a-94a9310fc952-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.713540 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" event={"ID":"9651259d-da1a-4580-855a-94a9310fc952","Type":"ContainerDied","Data":"75f857f14624805e9a86d248a55fa49b493451fc52a0d45c2fe39e01d81f1529"} Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.713587 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75f857f14624805e9a86d248a55fa49b493451fc52a0d45c2fe39e01d81f1529" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.713645 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w44w5" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.831112 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl"] Nov 24 17:44:17 crc kubenswrapper[4777]: E1124 17:44:17.831595 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d1f3579-f405-47c7-9728-f0a8508d88ac" containerName="extract-content" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.831614 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d1f3579-f405-47c7-9728-f0a8508d88ac" containerName="extract-content" Nov 24 17:44:17 crc kubenswrapper[4777]: E1124 17:44:17.831634 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d1f3579-f405-47c7-9728-f0a8508d88ac" containerName="extract-utilities" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.831644 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d1f3579-f405-47c7-9728-f0a8508d88ac" containerName="extract-utilities" Nov 24 17:44:17 crc kubenswrapper[4777]: E1124 17:44:17.831666 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d1f3579-f405-47c7-9728-f0a8508d88ac" containerName="registry-server" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.831674 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d1f3579-f405-47c7-9728-f0a8508d88ac" containerName="registry-server" Nov 24 17:44:17 crc kubenswrapper[4777]: E1124 17:44:17.831696 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9651259d-da1a-4580-855a-94a9310fc952" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.831704 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="9651259d-da1a-4580-855a-94a9310fc952" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.831947 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d1f3579-f405-47c7-9728-f0a8508d88ac" containerName="registry-server" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.832039 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="9651259d-da1a-4580-855a-94a9310fc952" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.833014 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.841419 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.841556 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-69hlj" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.841732 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.841847 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.841870 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 17:44:17 crc kubenswrapper[4777]: I1124 17:44:17.842748 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl"] Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.022375 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.022441 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.022549 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.022693 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89b75\" (UniqueName: \"kubernetes.io/projected/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-kube-api-access-89b75\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.022934 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.023004 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.023134 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.124805 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89b75\" (UniqueName: \"kubernetes.io/projected/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-kube-api-access-89b75\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.124916 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.124939 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.125006 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.125098 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.125132 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.125164 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.128645 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.128955 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.129396 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.130033 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.130482 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.130952 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.153039 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89b75\" (UniqueName: \"kubernetes.io/projected/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-kube-api-access-89b75\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.153586 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:44:18 crc kubenswrapper[4777]: I1124 17:44:18.710516 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl"] Nov 24 17:44:19 crc kubenswrapper[4777]: I1124 17:44:19.737169 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" event={"ID":"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e","Type":"ContainerStarted","Data":"97ec61716d6002943aff8e0d4c87e01b9d7e0125b2326abeeafb23936eb5d698"} Nov 24 17:44:19 crc kubenswrapper[4777]: I1124 17:44:19.737462 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" event={"ID":"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e","Type":"ContainerStarted","Data":"b8a6e290993bf8565f98d73ecc41805e05943f6127715989486c1b02094f16c7"} Nov 24 17:44:19 crc kubenswrapper[4777]: I1124 17:44:19.767374 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" podStartSLOduration=2.3784542699999998 podStartE2EDuration="2.767353535s" podCreationTimestamp="2025-11-24 17:44:17 +0000 UTC" firstStartedPulling="2025-11-24 17:44:18.724591399 +0000 UTC m=+2616.883786468" lastFinishedPulling="2025-11-24 17:44:19.113490654 +0000 UTC m=+2617.272685733" observedRunningTime="2025-11-24 17:44:19.753695427 +0000 UTC m=+2617.912890516" watchObservedRunningTime="2025-11-24 17:44:19.767353535 +0000 UTC m=+2617.926548584" Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.166957 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh"] Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.171239 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.175662 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.175699 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.187122 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh"] Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.305187 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-config-volume\") pod \"collect-profiles-29400105-pzchh\" (UID: \"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.305253 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-secret-volume\") pod \"collect-profiles-29400105-pzchh\" (UID: \"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.305416 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwlmd\" (UniqueName: \"kubernetes.io/projected/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-kube-api-access-kwlmd\") pod \"collect-profiles-29400105-pzchh\" (UID: \"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.407324 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwlmd\" (UniqueName: \"kubernetes.io/projected/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-kube-api-access-kwlmd\") pod \"collect-profiles-29400105-pzchh\" (UID: \"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.407485 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-config-volume\") pod \"collect-profiles-29400105-pzchh\" (UID: \"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.407510 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-secret-volume\") pod \"collect-profiles-29400105-pzchh\" (UID: \"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.409339 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-config-volume\") pod \"collect-profiles-29400105-pzchh\" (UID: \"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.421955 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-secret-volume\") pod \"collect-profiles-29400105-pzchh\" (UID: \"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.426686 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwlmd\" (UniqueName: \"kubernetes.io/projected/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-kube-api-access-kwlmd\") pod \"collect-profiles-29400105-pzchh\" (UID: \"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.512407 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" Nov 24 17:45:00 crc kubenswrapper[4777]: I1124 17:45:00.993249 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh"] Nov 24 17:45:01 crc kubenswrapper[4777]: I1124 17:45:01.240932 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" event={"ID":"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34","Type":"ContainerStarted","Data":"89aa895bf8000aa9041cc561e0cfaea23037c6c48d4e23fa622c69ee87e1cea0"} Nov 24 17:45:01 crc kubenswrapper[4777]: I1124 17:45:01.241286 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" event={"ID":"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34","Type":"ContainerStarted","Data":"486212eea651969bd90737b1f08ad983e6e92cbc7ab1939773928597f831c6ac"} Nov 24 17:45:01 crc kubenswrapper[4777]: I1124 17:45:01.266523 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" podStartSLOduration=1.26650108 podStartE2EDuration="1.26650108s" podCreationTimestamp="2025-11-24 17:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:45:01.263489184 +0000 UTC m=+2659.422684253" watchObservedRunningTime="2025-11-24 17:45:01.26650108 +0000 UTC m=+2659.425696129" Nov 24 17:45:02 crc kubenswrapper[4777]: I1124 17:45:02.256709 4777 generic.go:334] "Generic (PLEG): container finished" podID="ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34" containerID="89aa895bf8000aa9041cc561e0cfaea23037c6c48d4e23fa622c69ee87e1cea0" exitCode=0 Nov 24 17:45:02 crc kubenswrapper[4777]: I1124 17:45:02.256840 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" event={"ID":"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34","Type":"ContainerDied","Data":"89aa895bf8000aa9041cc561e0cfaea23037c6c48d4e23fa622c69ee87e1cea0"} Nov 24 17:45:03 crc kubenswrapper[4777]: I1124 17:45:03.730063 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" Nov 24 17:45:03 crc kubenswrapper[4777]: I1124 17:45:03.881739 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-config-volume\") pod \"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34\" (UID: \"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34\") " Nov 24 17:45:03 crc kubenswrapper[4777]: I1124 17:45:03.881789 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-secret-volume\") pod \"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34\" (UID: \"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34\") " Nov 24 17:45:03 crc kubenswrapper[4777]: I1124 17:45:03.881884 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwlmd\" (UniqueName: \"kubernetes.io/projected/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-kube-api-access-kwlmd\") pod \"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34\" (UID: \"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34\") " Nov 24 17:45:03 crc kubenswrapper[4777]: I1124 17:45:03.883238 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-config-volume" (OuterVolumeSpecName: "config-volume") pod "ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34" (UID: "ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:45:03 crc kubenswrapper[4777]: I1124 17:45:03.890139 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-kube-api-access-kwlmd" (OuterVolumeSpecName: "kube-api-access-kwlmd") pod "ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34" (UID: "ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34"). InnerVolumeSpecName "kube-api-access-kwlmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:45:03 crc kubenswrapper[4777]: I1124 17:45:03.895077 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34" (UID: "ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:45:03 crc kubenswrapper[4777]: I1124 17:45:03.984755 4777 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:03 crc kubenswrapper[4777]: I1124 17:45:03.985285 4777 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:03 crc kubenswrapper[4777]: I1124 17:45:03.985491 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwlmd\" (UniqueName: \"kubernetes.io/projected/ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34-kube-api-access-kwlmd\") on node \"crc\" DevicePath \"\"" Nov 24 17:45:04 crc kubenswrapper[4777]: I1124 17:45:04.281575 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" event={"ID":"ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34","Type":"ContainerDied","Data":"486212eea651969bd90737b1f08ad983e6e92cbc7ab1939773928597f831c6ac"} Nov 24 17:45:04 crc kubenswrapper[4777]: I1124 17:45:04.281908 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="486212eea651969bd90737b1f08ad983e6e92cbc7ab1939773928597f831c6ac" Nov 24 17:45:04 crc kubenswrapper[4777]: I1124 17:45:04.281624 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400105-pzchh" Nov 24 17:45:04 crc kubenswrapper[4777]: I1124 17:45:04.347676 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw"] Nov 24 17:45:04 crc kubenswrapper[4777]: I1124 17:45:04.355760 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400060-j98hw"] Nov 24 17:45:05 crc kubenswrapper[4777]: I1124 17:45:05.259173 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ec3567f-1218-441b-b33d-054f0c6a2e2d" path="/var/lib/kubelet/pods/7ec3567f-1218-441b-b33d-054f0c6a2e2d/volumes" Nov 24 17:45:14 crc kubenswrapper[4777]: I1124 17:45:14.270556 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:45:14 crc kubenswrapper[4777]: I1124 17:45:14.271589 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:45:44 crc kubenswrapper[4777]: I1124 17:45:44.272452 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:45:44 crc kubenswrapper[4777]: I1124 17:45:44.273003 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:45:48 crc kubenswrapper[4777]: I1124 17:45:48.697750 4777 scope.go:117] "RemoveContainer" containerID="ff812fb720a45f50229f29dee56fb02172ef403278db9ed610613eae9089da73" Nov 24 17:46:14 crc kubenswrapper[4777]: I1124 17:46:14.270871 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:46:14 crc kubenswrapper[4777]: I1124 17:46:14.271405 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:46:14 crc kubenswrapper[4777]: I1124 17:46:14.271476 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:46:14 crc kubenswrapper[4777]: I1124 17:46:14.272669 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"535d5108efbbfad99f1fa015e190845054474327b1f12cffe94f88b32d9cf532"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:46:14 crc kubenswrapper[4777]: I1124 17:46:14.272746 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://535d5108efbbfad99f1fa015e190845054474327b1f12cffe94f88b32d9cf532" gracePeriod=600 Nov 24 17:46:15 crc kubenswrapper[4777]: I1124 17:46:15.066165 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="535d5108efbbfad99f1fa015e190845054474327b1f12cffe94f88b32d9cf532" exitCode=0 Nov 24 17:46:15 crc kubenswrapper[4777]: I1124 17:46:15.066266 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"535d5108efbbfad99f1fa015e190845054474327b1f12cffe94f88b32d9cf532"} Nov 24 17:46:15 crc kubenswrapper[4777]: I1124 17:46:15.066735 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9"} Nov 24 17:46:15 crc kubenswrapper[4777]: I1124 17:46:15.066761 4777 scope.go:117] "RemoveContainer" containerID="654299b430eeafb48acd351eaebf4f2e0c380a5084ab17f98310ee9cbf1e534d" Nov 24 17:46:32 crc kubenswrapper[4777]: I1124 17:46:32.262464 4777 generic.go:334] "Generic (PLEG): container finished" podID="bcd7d063-a7cc-4f12-8ba2-9e7578152a3e" containerID="97ec61716d6002943aff8e0d4c87e01b9d7e0125b2326abeeafb23936eb5d698" exitCode=0 Nov 24 17:46:32 crc kubenswrapper[4777]: I1124 17:46:32.262757 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" event={"ID":"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e","Type":"ContainerDied","Data":"97ec61716d6002943aff8e0d4c87e01b9d7e0125b2326abeeafb23936eb5d698"} Nov 24 17:46:33 crc kubenswrapper[4777]: I1124 17:46:33.892532 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:46:33 crc kubenswrapper[4777]: I1124 17:46:33.996304 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-telemetry-combined-ca-bundle\") pod \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " Nov 24 17:46:33 crc kubenswrapper[4777]: I1124 17:46:33.996497 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-2\") pod \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " Nov 24 17:46:33 crc kubenswrapper[4777]: I1124 17:46:33.996563 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-inventory\") pod \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " Nov 24 17:46:33 crc kubenswrapper[4777]: I1124 17:46:33.996605 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-1\") pod \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " Nov 24 17:46:33 crc kubenswrapper[4777]: I1124 17:46:33.996642 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89b75\" (UniqueName: \"kubernetes.io/projected/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-kube-api-access-89b75\") pod \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " Nov 24 17:46:33 crc kubenswrapper[4777]: I1124 17:46:33.996689 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-0\") pod \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " Nov 24 17:46:33 crc kubenswrapper[4777]: I1124 17:46:33.996803 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ssh-key\") pod \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\" (UID: \"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e\") " Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.008318 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-kube-api-access-89b75" (OuterVolumeSpecName: "kube-api-access-89b75") pod "bcd7d063-a7cc-4f12-8ba2-9e7578152a3e" (UID: "bcd7d063-a7cc-4f12-8ba2-9e7578152a3e"). InnerVolumeSpecName "kube-api-access-89b75". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.013612 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "bcd7d063-a7cc-4f12-8ba2-9e7578152a3e" (UID: "bcd7d063-a7cc-4f12-8ba2-9e7578152a3e"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.027660 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "bcd7d063-a7cc-4f12-8ba2-9e7578152a3e" (UID: "bcd7d063-a7cc-4f12-8ba2-9e7578152a3e"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.031307 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "bcd7d063-a7cc-4f12-8ba2-9e7578152a3e" (UID: "bcd7d063-a7cc-4f12-8ba2-9e7578152a3e"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.033636 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bcd7d063-a7cc-4f12-8ba2-9e7578152a3e" (UID: "bcd7d063-a7cc-4f12-8ba2-9e7578152a3e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.039167 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-inventory" (OuterVolumeSpecName: "inventory") pod "bcd7d063-a7cc-4f12-8ba2-9e7578152a3e" (UID: "bcd7d063-a7cc-4f12-8ba2-9e7578152a3e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.045182 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "bcd7d063-a7cc-4f12-8ba2-9e7578152a3e" (UID: "bcd7d063-a7cc-4f12-8ba2-9e7578152a3e"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.099594 4777 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.099940 4777 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.099955 4777 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.100017 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89b75\" (UniqueName: \"kubernetes.io/projected/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-kube-api-access-89b75\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.100034 4777 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.100045 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.100053 4777 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcd7d063-a7cc-4f12-8ba2-9e7578152a3e-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.286759 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" event={"ID":"bcd7d063-a7cc-4f12-8ba2-9e7578152a3e","Type":"ContainerDied","Data":"b8a6e290993bf8565f98d73ecc41805e05943f6127715989486c1b02094f16c7"} Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.286807 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8a6e290993bf8565f98d73ecc41805e05943f6127715989486c1b02094f16c7" Nov 24 17:46:34 crc kubenswrapper[4777]: I1124 17:46:34.286828 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl" Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.753395 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7h4q8"] Nov 24 17:46:40 crc kubenswrapper[4777]: E1124 17:46:40.754686 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34" containerName="collect-profiles" Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.754711 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34" containerName="collect-profiles" Nov 24 17:46:40 crc kubenswrapper[4777]: E1124 17:46:40.754726 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcd7d063-a7cc-4f12-8ba2-9e7578152a3e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.754738 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcd7d063-a7cc-4f12-8ba2-9e7578152a3e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.755184 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcd7d063-a7cc-4f12-8ba2-9e7578152a3e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.755215 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccfc5bf6-6cb3-447d-b0a7-ed6b56c5be34" containerName="collect-profiles" Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.759731 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.773454 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7h4q8"] Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.861422 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-catalog-content\") pod \"community-operators-7h4q8\" (UID: \"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9\") " pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.861574 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txdq7\" (UniqueName: \"kubernetes.io/projected/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-kube-api-access-txdq7\") pod \"community-operators-7h4q8\" (UID: \"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9\") " pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.861622 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-utilities\") pod \"community-operators-7h4q8\" (UID: \"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9\") " pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.963683 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-catalog-content\") pod \"community-operators-7h4q8\" (UID: \"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9\") " pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.963818 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txdq7\" (UniqueName: \"kubernetes.io/projected/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-kube-api-access-txdq7\") pod \"community-operators-7h4q8\" (UID: \"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9\") " pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.963863 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-utilities\") pod \"community-operators-7h4q8\" (UID: \"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9\") " pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.964579 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-catalog-content\") pod \"community-operators-7h4q8\" (UID: \"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9\") " pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.964613 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-utilities\") pod \"community-operators-7h4q8\" (UID: \"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9\") " pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:40 crc kubenswrapper[4777]: I1124 17:46:40.987979 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txdq7\" (UniqueName: \"kubernetes.io/projected/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-kube-api-access-txdq7\") pod \"community-operators-7h4q8\" (UID: \"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9\") " pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:41 crc kubenswrapper[4777]: I1124 17:46:41.101458 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:41 crc kubenswrapper[4777]: I1124 17:46:41.612265 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7h4q8"] Nov 24 17:46:42 crc kubenswrapper[4777]: I1124 17:46:42.381939 4777 generic.go:334] "Generic (PLEG): container finished" podID="97cf68dc-9c01-4dc9-a2ad-cf09a52218b9" containerID="7f5d049ba5ba866fc97cee964f0eb66599e388201cc7a12490730457934e0972" exitCode=0 Nov 24 17:46:42 crc kubenswrapper[4777]: I1124 17:46:42.382087 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7h4q8" event={"ID":"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9","Type":"ContainerDied","Data":"7f5d049ba5ba866fc97cee964f0eb66599e388201cc7a12490730457934e0972"} Nov 24 17:46:42 crc kubenswrapper[4777]: I1124 17:46:42.382378 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7h4q8" event={"ID":"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9","Type":"ContainerStarted","Data":"333d3fe8e90fa44cc5d108cfedc76737583893e3d35957b41b1804b0395da204"} Nov 24 17:46:42 crc kubenswrapper[4777]: I1124 17:46:42.386105 4777 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:46:44 crc kubenswrapper[4777]: I1124 17:46:44.410044 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7h4q8" event={"ID":"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9","Type":"ContainerStarted","Data":"da498b60b536e95b8c2931d1bae1b3e23f1b15a7b54d1f16be6619f9a3249cbc"} Nov 24 17:46:45 crc kubenswrapper[4777]: I1124 17:46:45.424308 4777 generic.go:334] "Generic (PLEG): container finished" podID="97cf68dc-9c01-4dc9-a2ad-cf09a52218b9" containerID="da498b60b536e95b8c2931d1bae1b3e23f1b15a7b54d1f16be6619f9a3249cbc" exitCode=0 Nov 24 17:46:45 crc kubenswrapper[4777]: I1124 17:46:45.424377 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7h4q8" event={"ID":"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9","Type":"ContainerDied","Data":"da498b60b536e95b8c2931d1bae1b3e23f1b15a7b54d1f16be6619f9a3249cbc"} Nov 24 17:46:46 crc kubenswrapper[4777]: I1124 17:46:46.447306 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7h4q8" event={"ID":"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9","Type":"ContainerStarted","Data":"ce37b414120006ed4ba0dedf9ef1ac17c25bf5994dd7fa3b1ba566aa4b82e3a3"} Nov 24 17:46:46 crc kubenswrapper[4777]: I1124 17:46:46.482345 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7h4q8" podStartSLOduration=3.046683542 podStartE2EDuration="6.482318558s" podCreationTimestamp="2025-11-24 17:46:40 +0000 UTC" firstStartedPulling="2025-11-24 17:46:42.38573135 +0000 UTC m=+2760.544926409" lastFinishedPulling="2025-11-24 17:46:45.821366366 +0000 UTC m=+2763.980561425" observedRunningTime="2025-11-24 17:46:46.473746375 +0000 UTC m=+2764.632941444" watchObservedRunningTime="2025-11-24 17:46:46.482318558 +0000 UTC m=+2764.641513637" Nov 24 17:46:51 crc kubenswrapper[4777]: I1124 17:46:51.102557 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:51 crc kubenswrapper[4777]: I1124 17:46:51.103398 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:51 crc kubenswrapper[4777]: I1124 17:46:51.196421 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:51 crc kubenswrapper[4777]: I1124 17:46:51.585254 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:51 crc kubenswrapper[4777]: I1124 17:46:51.648330 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7h4q8"] Nov 24 17:46:53 crc kubenswrapper[4777]: I1124 17:46:53.530625 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7h4q8" podUID="97cf68dc-9c01-4dc9-a2ad-cf09a52218b9" containerName="registry-server" containerID="cri-o://ce37b414120006ed4ba0dedf9ef1ac17c25bf5994dd7fa3b1ba566aa4b82e3a3" gracePeriod=2 Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.076492 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.159739 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txdq7\" (UniqueName: \"kubernetes.io/projected/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-kube-api-access-txdq7\") pod \"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9\" (UID: \"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9\") " Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.160011 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-catalog-content\") pod \"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9\" (UID: \"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9\") " Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.160248 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-utilities\") pod \"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9\" (UID: \"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9\") " Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.163717 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-utilities" (OuterVolumeSpecName: "utilities") pod "97cf68dc-9c01-4dc9-a2ad-cf09a52218b9" (UID: "97cf68dc-9c01-4dc9-a2ad-cf09a52218b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.167867 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-kube-api-access-txdq7" (OuterVolumeSpecName: "kube-api-access-txdq7") pod "97cf68dc-9c01-4dc9-a2ad-cf09a52218b9" (UID: "97cf68dc-9c01-4dc9-a2ad-cf09a52218b9"). InnerVolumeSpecName "kube-api-access-txdq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.216284 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97cf68dc-9c01-4dc9-a2ad-cf09a52218b9" (UID: "97cf68dc-9c01-4dc9-a2ad-cf09a52218b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.262643 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.262690 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txdq7\" (UniqueName: \"kubernetes.io/projected/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-kube-api-access-txdq7\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.262707 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.546802 4777 generic.go:334] "Generic (PLEG): container finished" podID="97cf68dc-9c01-4dc9-a2ad-cf09a52218b9" containerID="ce37b414120006ed4ba0dedf9ef1ac17c25bf5994dd7fa3b1ba566aa4b82e3a3" exitCode=0 Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.546871 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7h4q8" event={"ID":"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9","Type":"ContainerDied","Data":"ce37b414120006ed4ba0dedf9ef1ac17c25bf5994dd7fa3b1ba566aa4b82e3a3"} Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.546904 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7h4q8" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.546949 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7h4q8" event={"ID":"97cf68dc-9c01-4dc9-a2ad-cf09a52218b9","Type":"ContainerDied","Data":"333d3fe8e90fa44cc5d108cfedc76737583893e3d35957b41b1804b0395da204"} Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.546999 4777 scope.go:117] "RemoveContainer" containerID="ce37b414120006ed4ba0dedf9ef1ac17c25bf5994dd7fa3b1ba566aa4b82e3a3" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.574442 4777 scope.go:117] "RemoveContainer" containerID="da498b60b536e95b8c2931d1bae1b3e23f1b15a7b54d1f16be6619f9a3249cbc" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.614286 4777 scope.go:117] "RemoveContainer" containerID="7f5d049ba5ba866fc97cee964f0eb66599e388201cc7a12490730457934e0972" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.620130 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7h4q8"] Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.631846 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7h4q8"] Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.673818 4777 scope.go:117] "RemoveContainer" containerID="ce37b414120006ed4ba0dedf9ef1ac17c25bf5994dd7fa3b1ba566aa4b82e3a3" Nov 24 17:46:54 crc kubenswrapper[4777]: E1124 17:46:54.674418 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce37b414120006ed4ba0dedf9ef1ac17c25bf5994dd7fa3b1ba566aa4b82e3a3\": container with ID starting with ce37b414120006ed4ba0dedf9ef1ac17c25bf5994dd7fa3b1ba566aa4b82e3a3 not found: ID does not exist" containerID="ce37b414120006ed4ba0dedf9ef1ac17c25bf5994dd7fa3b1ba566aa4b82e3a3" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.674457 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce37b414120006ed4ba0dedf9ef1ac17c25bf5994dd7fa3b1ba566aa4b82e3a3"} err="failed to get container status \"ce37b414120006ed4ba0dedf9ef1ac17c25bf5994dd7fa3b1ba566aa4b82e3a3\": rpc error: code = NotFound desc = could not find container \"ce37b414120006ed4ba0dedf9ef1ac17c25bf5994dd7fa3b1ba566aa4b82e3a3\": container with ID starting with ce37b414120006ed4ba0dedf9ef1ac17c25bf5994dd7fa3b1ba566aa4b82e3a3 not found: ID does not exist" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.674485 4777 scope.go:117] "RemoveContainer" containerID="da498b60b536e95b8c2931d1bae1b3e23f1b15a7b54d1f16be6619f9a3249cbc" Nov 24 17:46:54 crc kubenswrapper[4777]: E1124 17:46:54.674811 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da498b60b536e95b8c2931d1bae1b3e23f1b15a7b54d1f16be6619f9a3249cbc\": container with ID starting with da498b60b536e95b8c2931d1bae1b3e23f1b15a7b54d1f16be6619f9a3249cbc not found: ID does not exist" containerID="da498b60b536e95b8c2931d1bae1b3e23f1b15a7b54d1f16be6619f9a3249cbc" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.674844 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da498b60b536e95b8c2931d1bae1b3e23f1b15a7b54d1f16be6619f9a3249cbc"} err="failed to get container status \"da498b60b536e95b8c2931d1bae1b3e23f1b15a7b54d1f16be6619f9a3249cbc\": rpc error: code = NotFound desc = could not find container \"da498b60b536e95b8c2931d1bae1b3e23f1b15a7b54d1f16be6619f9a3249cbc\": container with ID starting with da498b60b536e95b8c2931d1bae1b3e23f1b15a7b54d1f16be6619f9a3249cbc not found: ID does not exist" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.674867 4777 scope.go:117] "RemoveContainer" containerID="7f5d049ba5ba866fc97cee964f0eb66599e388201cc7a12490730457934e0972" Nov 24 17:46:54 crc kubenswrapper[4777]: E1124 17:46:54.675240 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f5d049ba5ba866fc97cee964f0eb66599e388201cc7a12490730457934e0972\": container with ID starting with 7f5d049ba5ba866fc97cee964f0eb66599e388201cc7a12490730457934e0972 not found: ID does not exist" containerID="7f5d049ba5ba866fc97cee964f0eb66599e388201cc7a12490730457934e0972" Nov 24 17:46:54 crc kubenswrapper[4777]: I1124 17:46:54.675271 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f5d049ba5ba866fc97cee964f0eb66599e388201cc7a12490730457934e0972"} err="failed to get container status \"7f5d049ba5ba866fc97cee964f0eb66599e388201cc7a12490730457934e0972\": rpc error: code = NotFound desc = could not find container \"7f5d049ba5ba866fc97cee964f0eb66599e388201cc7a12490730457934e0972\": container with ID starting with 7f5d049ba5ba866fc97cee964f0eb66599e388201cc7a12490730457934e0972 not found: ID does not exist" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.268634 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97cf68dc-9c01-4dc9-a2ad-cf09a52218b9" path="/var/lib/kubelet/pods/97cf68dc-9c01-4dc9-a2ad-cf09a52218b9/volumes" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.673026 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6gwrh"] Nov 24 17:46:55 crc kubenswrapper[4777]: E1124 17:46:55.674017 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97cf68dc-9c01-4dc9-a2ad-cf09a52218b9" containerName="extract-content" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.674050 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="97cf68dc-9c01-4dc9-a2ad-cf09a52218b9" containerName="extract-content" Nov 24 17:46:55 crc kubenswrapper[4777]: E1124 17:46:55.674097 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97cf68dc-9c01-4dc9-a2ad-cf09a52218b9" containerName="registry-server" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.674119 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="97cf68dc-9c01-4dc9-a2ad-cf09a52218b9" containerName="registry-server" Nov 24 17:46:55 crc kubenswrapper[4777]: E1124 17:46:55.674159 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97cf68dc-9c01-4dc9-a2ad-cf09a52218b9" containerName="extract-utilities" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.674177 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="97cf68dc-9c01-4dc9-a2ad-cf09a52218b9" containerName="extract-utilities" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.674698 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="97cf68dc-9c01-4dc9-a2ad-cf09a52218b9" containerName="registry-server" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.684241 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.713014 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0718e52a-7534-496b-bf88-a9211606217e-catalog-content\") pod \"redhat-marketplace-6gwrh\" (UID: \"0718e52a-7534-496b-bf88-a9211606217e\") " pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.713596 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6gwrh"] Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.717072 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0718e52a-7534-496b-bf88-a9211606217e-utilities\") pod \"redhat-marketplace-6gwrh\" (UID: \"0718e52a-7534-496b-bf88-a9211606217e\") " pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.717168 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6hcs\" (UniqueName: \"kubernetes.io/projected/0718e52a-7534-496b-bf88-a9211606217e-kube-api-access-n6hcs\") pod \"redhat-marketplace-6gwrh\" (UID: \"0718e52a-7534-496b-bf88-a9211606217e\") " pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.819069 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0718e52a-7534-496b-bf88-a9211606217e-catalog-content\") pod \"redhat-marketplace-6gwrh\" (UID: \"0718e52a-7534-496b-bf88-a9211606217e\") " pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.819133 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0718e52a-7534-496b-bf88-a9211606217e-utilities\") pod \"redhat-marketplace-6gwrh\" (UID: \"0718e52a-7534-496b-bf88-a9211606217e\") " pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.819161 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6hcs\" (UniqueName: \"kubernetes.io/projected/0718e52a-7534-496b-bf88-a9211606217e-kube-api-access-n6hcs\") pod \"redhat-marketplace-6gwrh\" (UID: \"0718e52a-7534-496b-bf88-a9211606217e\") " pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.819558 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0718e52a-7534-496b-bf88-a9211606217e-catalog-content\") pod \"redhat-marketplace-6gwrh\" (UID: \"0718e52a-7534-496b-bf88-a9211606217e\") " pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.819909 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0718e52a-7534-496b-bf88-a9211606217e-utilities\") pod \"redhat-marketplace-6gwrh\" (UID: \"0718e52a-7534-496b-bf88-a9211606217e\") " pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:46:55 crc kubenswrapper[4777]: I1124 17:46:55.840412 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6hcs\" (UniqueName: \"kubernetes.io/projected/0718e52a-7534-496b-bf88-a9211606217e-kube-api-access-n6hcs\") pod \"redhat-marketplace-6gwrh\" (UID: \"0718e52a-7534-496b-bf88-a9211606217e\") " pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:46:56 crc kubenswrapper[4777]: I1124 17:46:56.030053 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:46:56 crc kubenswrapper[4777]: I1124 17:46:56.512674 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6gwrh"] Nov 24 17:46:56 crc kubenswrapper[4777]: W1124 17:46:56.521426 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0718e52a_7534_496b_bf88_a9211606217e.slice/crio-8cf556052461f2bf2abfdadd057a646b9addc12cc0c1bed91ed8345950e048bc WatchSource:0}: Error finding container 8cf556052461f2bf2abfdadd057a646b9addc12cc0c1bed91ed8345950e048bc: Status 404 returned error can't find the container with id 8cf556052461f2bf2abfdadd057a646b9addc12cc0c1bed91ed8345950e048bc Nov 24 17:46:56 crc kubenswrapper[4777]: I1124 17:46:56.574283 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6gwrh" event={"ID":"0718e52a-7534-496b-bf88-a9211606217e","Type":"ContainerStarted","Data":"8cf556052461f2bf2abfdadd057a646b9addc12cc0c1bed91ed8345950e048bc"} Nov 24 17:46:57 crc kubenswrapper[4777]: I1124 17:46:57.591177 4777 generic.go:334] "Generic (PLEG): container finished" podID="0718e52a-7534-496b-bf88-a9211606217e" containerID="eb4cde4342373c060014985d10b7ca0f929911fe70a029d7979b1c3f9583bdd4" exitCode=0 Nov 24 17:46:57 crc kubenswrapper[4777]: I1124 17:46:57.591250 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6gwrh" event={"ID":"0718e52a-7534-496b-bf88-a9211606217e","Type":"ContainerDied","Data":"eb4cde4342373c060014985d10b7ca0f929911fe70a029d7979b1c3f9583bdd4"} Nov 24 17:47:01 crc kubenswrapper[4777]: I1124 17:47:01.641605 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6gwrh" event={"ID":"0718e52a-7534-496b-bf88-a9211606217e","Type":"ContainerStarted","Data":"c87c788d9ef017ce774b49dcbe4a27d1d3d00b95a030d7308b9ca284b781a69b"} Nov 24 17:47:02 crc kubenswrapper[4777]: I1124 17:47:02.657025 4777 generic.go:334] "Generic (PLEG): container finished" podID="0718e52a-7534-496b-bf88-a9211606217e" containerID="c87c788d9ef017ce774b49dcbe4a27d1d3d00b95a030d7308b9ca284b781a69b" exitCode=0 Nov 24 17:47:02 crc kubenswrapper[4777]: I1124 17:47:02.657076 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6gwrh" event={"ID":"0718e52a-7534-496b-bf88-a9211606217e","Type":"ContainerDied","Data":"c87c788d9ef017ce774b49dcbe4a27d1d3d00b95a030d7308b9ca284b781a69b"} Nov 24 17:47:03 crc kubenswrapper[4777]: I1124 17:47:03.672518 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6gwrh" event={"ID":"0718e52a-7534-496b-bf88-a9211606217e","Type":"ContainerStarted","Data":"ebcb29a869f5040128ad187908e7c0a9efaa0d5220467120b024895d92c8133a"} Nov 24 17:47:03 crc kubenswrapper[4777]: I1124 17:47:03.701883 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6gwrh" podStartSLOduration=3.196982475 podStartE2EDuration="8.701862937s" podCreationTimestamp="2025-11-24 17:46:55 +0000 UTC" firstStartedPulling="2025-11-24 17:46:57.593505769 +0000 UTC m=+2775.752700818" lastFinishedPulling="2025-11-24 17:47:03.098386221 +0000 UTC m=+2781.257581280" observedRunningTime="2025-11-24 17:47:03.693992992 +0000 UTC m=+2781.853188051" watchObservedRunningTime="2025-11-24 17:47:03.701862937 +0000 UTC m=+2781.861057996" Nov 24 17:47:06 crc kubenswrapper[4777]: I1124 17:47:06.030681 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:47:06 crc kubenswrapper[4777]: I1124 17:47:06.031179 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:47:06 crc kubenswrapper[4777]: I1124 17:47:06.088838 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:47:16 crc kubenswrapper[4777]: I1124 17:47:16.108185 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:47:16 crc kubenswrapper[4777]: I1124 17:47:16.160875 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6gwrh"] Nov 24 17:47:16 crc kubenswrapper[4777]: I1124 17:47:16.874952 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6gwrh" podUID="0718e52a-7534-496b-bf88-a9211606217e" containerName="registry-server" containerID="cri-o://ebcb29a869f5040128ad187908e7c0a9efaa0d5220467120b024895d92c8133a" gracePeriod=2 Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.453124 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.621791 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6hcs\" (UniqueName: \"kubernetes.io/projected/0718e52a-7534-496b-bf88-a9211606217e-kube-api-access-n6hcs\") pod \"0718e52a-7534-496b-bf88-a9211606217e\" (UID: \"0718e52a-7534-496b-bf88-a9211606217e\") " Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.622070 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0718e52a-7534-496b-bf88-a9211606217e-utilities\") pod \"0718e52a-7534-496b-bf88-a9211606217e\" (UID: \"0718e52a-7534-496b-bf88-a9211606217e\") " Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.622208 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0718e52a-7534-496b-bf88-a9211606217e-catalog-content\") pod \"0718e52a-7534-496b-bf88-a9211606217e\" (UID: \"0718e52a-7534-496b-bf88-a9211606217e\") " Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.624463 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0718e52a-7534-496b-bf88-a9211606217e-utilities" (OuterVolumeSpecName: "utilities") pod "0718e52a-7534-496b-bf88-a9211606217e" (UID: "0718e52a-7534-496b-bf88-a9211606217e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.630073 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0718e52a-7534-496b-bf88-a9211606217e-kube-api-access-n6hcs" (OuterVolumeSpecName: "kube-api-access-n6hcs") pod "0718e52a-7534-496b-bf88-a9211606217e" (UID: "0718e52a-7534-496b-bf88-a9211606217e"). InnerVolumeSpecName "kube-api-access-n6hcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.642201 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0718e52a-7534-496b-bf88-a9211606217e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0718e52a-7534-496b-bf88-a9211606217e" (UID: "0718e52a-7534-496b-bf88-a9211606217e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.724779 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0718e52a-7534-496b-bf88-a9211606217e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.724820 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0718e52a-7534-496b-bf88-a9211606217e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.724839 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6hcs\" (UniqueName: \"kubernetes.io/projected/0718e52a-7534-496b-bf88-a9211606217e-kube-api-access-n6hcs\") on node \"crc\" DevicePath \"\"" Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.893330 4777 generic.go:334] "Generic (PLEG): container finished" podID="0718e52a-7534-496b-bf88-a9211606217e" containerID="ebcb29a869f5040128ad187908e7c0a9efaa0d5220467120b024895d92c8133a" exitCode=0 Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.893369 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6gwrh" event={"ID":"0718e52a-7534-496b-bf88-a9211606217e","Type":"ContainerDied","Data":"ebcb29a869f5040128ad187908e7c0a9efaa0d5220467120b024895d92c8133a"} Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.893423 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6gwrh" event={"ID":"0718e52a-7534-496b-bf88-a9211606217e","Type":"ContainerDied","Data":"8cf556052461f2bf2abfdadd057a646b9addc12cc0c1bed91ed8345950e048bc"} Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.893451 4777 scope.go:117] "RemoveContainer" containerID="ebcb29a869f5040128ad187908e7c0a9efaa0d5220467120b024895d92c8133a" Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.894218 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6gwrh" Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.930053 4777 scope.go:117] "RemoveContainer" containerID="c87c788d9ef017ce774b49dcbe4a27d1d3d00b95a030d7308b9ca284b781a69b" Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.952381 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6gwrh"] Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.965435 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6gwrh"] Nov 24 17:47:17 crc kubenswrapper[4777]: I1124 17:47:17.966119 4777 scope.go:117] "RemoveContainer" containerID="eb4cde4342373c060014985d10b7ca0f929911fe70a029d7979b1c3f9583bdd4" Nov 24 17:47:18 crc kubenswrapper[4777]: I1124 17:47:18.012154 4777 scope.go:117] "RemoveContainer" containerID="ebcb29a869f5040128ad187908e7c0a9efaa0d5220467120b024895d92c8133a" Nov 24 17:47:18 crc kubenswrapper[4777]: E1124 17:47:18.012594 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebcb29a869f5040128ad187908e7c0a9efaa0d5220467120b024895d92c8133a\": container with ID starting with ebcb29a869f5040128ad187908e7c0a9efaa0d5220467120b024895d92c8133a not found: ID does not exist" containerID="ebcb29a869f5040128ad187908e7c0a9efaa0d5220467120b024895d92c8133a" Nov 24 17:47:18 crc kubenswrapper[4777]: I1124 17:47:18.012683 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebcb29a869f5040128ad187908e7c0a9efaa0d5220467120b024895d92c8133a"} err="failed to get container status \"ebcb29a869f5040128ad187908e7c0a9efaa0d5220467120b024895d92c8133a\": rpc error: code = NotFound desc = could not find container \"ebcb29a869f5040128ad187908e7c0a9efaa0d5220467120b024895d92c8133a\": container with ID starting with ebcb29a869f5040128ad187908e7c0a9efaa0d5220467120b024895d92c8133a not found: ID does not exist" Nov 24 17:47:18 crc kubenswrapper[4777]: I1124 17:47:18.012710 4777 scope.go:117] "RemoveContainer" containerID="c87c788d9ef017ce774b49dcbe4a27d1d3d00b95a030d7308b9ca284b781a69b" Nov 24 17:47:18 crc kubenswrapper[4777]: E1124 17:47:18.013159 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c87c788d9ef017ce774b49dcbe4a27d1d3d00b95a030d7308b9ca284b781a69b\": container with ID starting with c87c788d9ef017ce774b49dcbe4a27d1d3d00b95a030d7308b9ca284b781a69b not found: ID does not exist" containerID="c87c788d9ef017ce774b49dcbe4a27d1d3d00b95a030d7308b9ca284b781a69b" Nov 24 17:47:18 crc kubenswrapper[4777]: I1124 17:47:18.013188 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c87c788d9ef017ce774b49dcbe4a27d1d3d00b95a030d7308b9ca284b781a69b"} err="failed to get container status \"c87c788d9ef017ce774b49dcbe4a27d1d3d00b95a030d7308b9ca284b781a69b\": rpc error: code = NotFound desc = could not find container \"c87c788d9ef017ce774b49dcbe4a27d1d3d00b95a030d7308b9ca284b781a69b\": container with ID starting with c87c788d9ef017ce774b49dcbe4a27d1d3d00b95a030d7308b9ca284b781a69b not found: ID does not exist" Nov 24 17:47:18 crc kubenswrapper[4777]: I1124 17:47:18.013201 4777 scope.go:117] "RemoveContainer" containerID="eb4cde4342373c060014985d10b7ca0f929911fe70a029d7979b1c3f9583bdd4" Nov 24 17:47:18 crc kubenswrapper[4777]: E1124 17:47:18.013594 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb4cde4342373c060014985d10b7ca0f929911fe70a029d7979b1c3f9583bdd4\": container with ID starting with eb4cde4342373c060014985d10b7ca0f929911fe70a029d7979b1c3f9583bdd4 not found: ID does not exist" containerID="eb4cde4342373c060014985d10b7ca0f929911fe70a029d7979b1c3f9583bdd4" Nov 24 17:47:18 crc kubenswrapper[4777]: I1124 17:47:18.013648 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb4cde4342373c060014985d10b7ca0f929911fe70a029d7979b1c3f9583bdd4"} err="failed to get container status \"eb4cde4342373c060014985d10b7ca0f929911fe70a029d7979b1c3f9583bdd4\": rpc error: code = NotFound desc = could not find container \"eb4cde4342373c060014985d10b7ca0f929911fe70a029d7979b1c3f9583bdd4\": container with ID starting with eb4cde4342373c060014985d10b7ca0f929911fe70a029d7979b1c3f9583bdd4 not found: ID does not exist" Nov 24 17:47:19 crc kubenswrapper[4777]: I1124 17:47:19.260865 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0718e52a-7534-496b-bf88-a9211606217e" path="/var/lib/kubelet/pods/0718e52a-7534-496b-bf88-a9211606217e/volumes" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.490528 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 17:47:27 crc kubenswrapper[4777]: E1124 17:47:27.491886 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0718e52a-7534-496b-bf88-a9211606217e" containerName="extract-utilities" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.491904 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="0718e52a-7534-496b-bf88-a9211606217e" containerName="extract-utilities" Nov 24 17:47:27 crc kubenswrapper[4777]: E1124 17:47:27.491936 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0718e52a-7534-496b-bf88-a9211606217e" containerName="registry-server" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.491944 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="0718e52a-7534-496b-bf88-a9211606217e" containerName="registry-server" Nov 24 17:47:27 crc kubenswrapper[4777]: E1124 17:47:27.491986 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0718e52a-7534-496b-bf88-a9211606217e" containerName="extract-content" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.491995 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="0718e52a-7534-496b-bf88-a9211606217e" containerName="extract-content" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.492264 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="0718e52a-7534-496b-bf88-a9211606217e" containerName="registry-server" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.493208 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.495832 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.496087 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-77z9z" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.496330 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.499259 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.520270 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.642212 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.642277 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n8p8\" (UniqueName: \"kubernetes.io/projected/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-kube-api-access-6n8p8\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.642317 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.642358 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.642465 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.642485 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-config-data\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.642551 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.642616 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.642649 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.744661 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.745122 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.745229 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.745278 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n8p8\" (UniqueName: \"kubernetes.io/projected/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-kube-api-access-6n8p8\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.745333 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.745390 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.745519 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.745562 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-config-data\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.745672 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.746153 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.747818 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.748755 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-config-data\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.749235 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.751352 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.752305 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.752992 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.760483 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.778625 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n8p8\" (UniqueName: \"kubernetes.io/projected/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-kube-api-access-6n8p8\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.798733 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " pod="openstack/tempest-tests-tempest" Nov 24 17:47:27 crc kubenswrapper[4777]: I1124 17:47:27.832432 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 17:47:28 crc kubenswrapper[4777]: I1124 17:47:28.301682 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 17:47:29 crc kubenswrapper[4777]: I1124 17:47:29.041769 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"08d58d9b-d4f1-458d-bfe3-fa8bc585f352","Type":"ContainerStarted","Data":"579737412bf8f8b1fa4126e41047c2a2ccaee6102ca29898ec540ef7021d966f"} Nov 24 17:48:01 crc kubenswrapper[4777]: E1124 17:48:01.068156 4777 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 24 17:48:01 crc kubenswrapper[4777]: E1124 17:48:01.071947 4777 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6n8p8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(08d58d9b-d4f1-458d-bfe3-fa8bc585f352): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 17:48:01 crc kubenswrapper[4777]: E1124 17:48:01.073746 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="08d58d9b-d4f1-458d-bfe3-fa8bc585f352" Nov 24 17:48:01 crc kubenswrapper[4777]: E1124 17:48:01.448383 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="08d58d9b-d4f1-458d-bfe3-fa8bc585f352" Nov 24 17:48:14 crc kubenswrapper[4777]: I1124 17:48:14.270724 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:48:14 crc kubenswrapper[4777]: I1124 17:48:14.272448 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:48:15 crc kubenswrapper[4777]: I1124 17:48:15.721401 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 24 17:48:17 crc kubenswrapper[4777]: I1124 17:48:17.634357 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"08d58d9b-d4f1-458d-bfe3-fa8bc585f352","Type":"ContainerStarted","Data":"b2d82616ac0616ec9cfeb48fd46f24961bd6b01e6c2dce49d459f3b3cb9eb916"} Nov 24 17:48:17 crc kubenswrapper[4777]: I1124 17:48:17.664488 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.259420816 podStartE2EDuration="51.664454521s" podCreationTimestamp="2025-11-24 17:47:26 +0000 UTC" firstStartedPulling="2025-11-24 17:47:28.313763838 +0000 UTC m=+2806.472958887" lastFinishedPulling="2025-11-24 17:48:15.718797543 +0000 UTC m=+2853.877992592" observedRunningTime="2025-11-24 17:48:17.657563033 +0000 UTC m=+2855.816758082" watchObservedRunningTime="2025-11-24 17:48:17.664454521 +0000 UTC m=+2855.823649610" Nov 24 17:48:44 crc kubenswrapper[4777]: I1124 17:48:44.271147 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:48:44 crc kubenswrapper[4777]: I1124 17:48:44.272200 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:49:14 crc kubenswrapper[4777]: I1124 17:49:14.270040 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:49:14 crc kubenswrapper[4777]: I1124 17:49:14.270595 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:49:14 crc kubenswrapper[4777]: I1124 17:49:14.270640 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:49:14 crc kubenswrapper[4777]: I1124 17:49:14.271486 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:49:14 crc kubenswrapper[4777]: I1124 17:49:14.271553 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" gracePeriod=600 Nov 24 17:49:14 crc kubenswrapper[4777]: E1124 17:49:14.399199 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:49:15 crc kubenswrapper[4777]: I1124 17:49:15.278875 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" exitCode=0 Nov 24 17:49:15 crc kubenswrapper[4777]: I1124 17:49:15.278962 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9"} Nov 24 17:49:15 crc kubenswrapper[4777]: I1124 17:49:15.279017 4777 scope.go:117] "RemoveContainer" containerID="535d5108efbbfad99f1fa015e190845054474327b1f12cffe94f88b32d9cf532" Nov 24 17:49:15 crc kubenswrapper[4777]: I1124 17:49:15.279852 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:49:15 crc kubenswrapper[4777]: E1124 17:49:15.280397 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:49:28 crc kubenswrapper[4777]: I1124 17:49:28.245462 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:49:28 crc kubenswrapper[4777]: E1124 17:49:28.246132 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:49:39 crc kubenswrapper[4777]: I1124 17:49:39.246033 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:49:39 crc kubenswrapper[4777]: E1124 17:49:39.246828 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:49:53 crc kubenswrapper[4777]: I1124 17:49:53.252223 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:49:53 crc kubenswrapper[4777]: E1124 17:49:53.253151 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:50:08 crc kubenswrapper[4777]: I1124 17:50:08.245748 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:50:08 crc kubenswrapper[4777]: E1124 17:50:08.246575 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:50:22 crc kubenswrapper[4777]: I1124 17:50:22.245308 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:50:22 crc kubenswrapper[4777]: E1124 17:50:22.246041 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:50:34 crc kubenswrapper[4777]: I1124 17:50:34.245351 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:50:34 crc kubenswrapper[4777]: E1124 17:50:34.246269 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:50:47 crc kubenswrapper[4777]: I1124 17:50:47.248863 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:50:47 crc kubenswrapper[4777]: E1124 17:50:47.250057 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:51:01 crc kubenswrapper[4777]: I1124 17:51:01.244981 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:51:01 crc kubenswrapper[4777]: E1124 17:51:01.245843 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:51:16 crc kubenswrapper[4777]: I1124 17:51:16.245789 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:51:16 crc kubenswrapper[4777]: E1124 17:51:16.246478 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:51:27 crc kubenswrapper[4777]: I1124 17:51:27.250274 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:51:27 crc kubenswrapper[4777]: E1124 17:51:27.251322 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:51:39 crc kubenswrapper[4777]: I1124 17:51:39.245773 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:51:39 crc kubenswrapper[4777]: E1124 17:51:39.246873 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:51:52 crc kubenswrapper[4777]: I1124 17:51:52.245416 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:51:52 crc kubenswrapper[4777]: E1124 17:51:52.246708 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:52:07 crc kubenswrapper[4777]: I1124 17:52:07.245295 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:52:07 crc kubenswrapper[4777]: E1124 17:52:07.246086 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:52:21 crc kubenswrapper[4777]: I1124 17:52:21.245665 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:52:21 crc kubenswrapper[4777]: E1124 17:52:21.246402 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:52:34 crc kubenswrapper[4777]: I1124 17:52:34.245131 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:52:34 crc kubenswrapper[4777]: E1124 17:52:34.246073 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:52:49 crc kubenswrapper[4777]: I1124 17:52:49.246105 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:52:49 crc kubenswrapper[4777]: E1124 17:52:49.246939 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:53:02 crc kubenswrapper[4777]: I1124 17:53:02.245460 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:53:02 crc kubenswrapper[4777]: E1124 17:53:02.246495 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:53:15 crc kubenswrapper[4777]: I1124 17:53:15.246590 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:53:15 crc kubenswrapper[4777]: E1124 17:53:15.247382 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:53:21 crc kubenswrapper[4777]: I1124 17:53:21.797774 4777 generic.go:334] "Generic (PLEG): container finished" podID="08d58d9b-d4f1-458d-bfe3-fa8bc585f352" containerID="b2d82616ac0616ec9cfeb48fd46f24961bd6b01e6c2dce49d459f3b3cb9eb916" exitCode=0 Nov 24 17:53:21 crc kubenswrapper[4777]: I1124 17:53:21.797876 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"08d58d9b-d4f1-458d-bfe3-fa8bc585f352","Type":"ContainerDied","Data":"b2d82616ac0616ec9cfeb48fd46f24961bd6b01e6c2dce49d459f3b3cb9eb916"} Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.496030 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.624279 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-test-operator-ephemeral-temporary\") pod \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.624313 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.624342 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-openstack-config\") pod \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.624392 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-ssh-key\") pod \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.624496 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-config-data\") pod \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.624535 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n8p8\" (UniqueName: \"kubernetes.io/projected/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-kube-api-access-6n8p8\") pod \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.624554 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-openstack-config-secret\") pod \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.625057 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "08d58d9b-d4f1-458d-bfe3-fa8bc585f352" (UID: "08d58d9b-d4f1-458d-bfe3-fa8bc585f352"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.625172 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-config-data" (OuterVolumeSpecName: "config-data") pod "08d58d9b-d4f1-458d-bfe3-fa8bc585f352" (UID: "08d58d9b-d4f1-458d-bfe3-fa8bc585f352"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.625312 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-ca-certs\") pod \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.625366 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-test-operator-ephemeral-workdir\") pod \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\" (UID: \"08d58d9b-d4f1-458d-bfe3-fa8bc585f352\") " Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.625857 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.625880 4777 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.660988 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "08d58d9b-d4f1-458d-bfe3-fa8bc585f352" (UID: "08d58d9b-d4f1-458d-bfe3-fa8bc585f352"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.661089 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-kube-api-access-6n8p8" (OuterVolumeSpecName: "kube-api-access-6n8p8") pod "08d58d9b-d4f1-458d-bfe3-fa8bc585f352" (UID: "08d58d9b-d4f1-458d-bfe3-fa8bc585f352"). InnerVolumeSpecName "kube-api-access-6n8p8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.671122 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "08d58d9b-d4f1-458d-bfe3-fa8bc585f352" (UID: "08d58d9b-d4f1-458d-bfe3-fa8bc585f352"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.672123 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "08d58d9b-d4f1-458d-bfe3-fa8bc585f352" (UID: "08d58d9b-d4f1-458d-bfe3-fa8bc585f352"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.698493 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "08d58d9b-d4f1-458d-bfe3-fa8bc585f352" (UID: "08d58d9b-d4f1-458d-bfe3-fa8bc585f352"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.731057 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n8p8\" (UniqueName: \"kubernetes.io/projected/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-kube-api-access-6n8p8\") on node \"crc\" DevicePath \"\"" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.731086 4777 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.731108 4777 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.731117 4777 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.731127 4777 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.788227 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "08d58d9b-d4f1-458d-bfe3-fa8bc585f352" (UID: "08d58d9b-d4f1-458d-bfe3-fa8bc585f352"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.793027 4777 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.823760 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"08d58d9b-d4f1-458d-bfe3-fa8bc585f352","Type":"ContainerDied","Data":"579737412bf8f8b1fa4126e41047c2a2ccaee6102ca29898ec540ef7021d966f"} Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.823863 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="579737412bf8f8b1fa4126e41047c2a2ccaee6102ca29898ec540ef7021d966f" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.824004 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.833199 4777 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 24 17:53:23 crc kubenswrapper[4777]: I1124 17:53:23.833461 4777 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 24 17:53:24 crc kubenswrapper[4777]: I1124 17:53:24.046558 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "08d58d9b-d4f1-458d-bfe3-fa8bc585f352" (UID: "08d58d9b-d4f1-458d-bfe3-fa8bc585f352"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:53:24 crc kubenswrapper[4777]: I1124 17:53:24.139032 4777 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/08d58d9b-d4f1-458d-bfe3-fa8bc585f352-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 24 17:53:26 crc kubenswrapper[4777]: I1124 17:53:26.708558 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 17:53:26 crc kubenswrapper[4777]: E1124 17:53:26.710108 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d58d9b-d4f1-458d-bfe3-fa8bc585f352" containerName="tempest-tests-tempest-tests-runner" Nov 24 17:53:26 crc kubenswrapper[4777]: I1124 17:53:26.710139 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d58d9b-d4f1-458d-bfe3-fa8bc585f352" containerName="tempest-tests-tempest-tests-runner" Nov 24 17:53:26 crc kubenswrapper[4777]: I1124 17:53:26.710587 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="08d58d9b-d4f1-458d-bfe3-fa8bc585f352" containerName="tempest-tests-tempest-tests-runner" Nov 24 17:53:26 crc kubenswrapper[4777]: I1124 17:53:26.713026 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 17:53:26 crc kubenswrapper[4777]: I1124 17:53:26.720349 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-77z9z" Nov 24 17:53:26 crc kubenswrapper[4777]: I1124 17:53:26.738098 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 17:53:26 crc kubenswrapper[4777]: I1124 17:53:26.793052 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b55r5\" (UniqueName: \"kubernetes.io/projected/2d149ad6-8d86-445c-baed-c43f1eaec2cb-kube-api-access-b55r5\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2d149ad6-8d86-445c-baed-c43f1eaec2cb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 17:53:26 crc kubenswrapper[4777]: I1124 17:53:26.793225 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2d149ad6-8d86-445c-baed-c43f1eaec2cb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 17:53:26 crc kubenswrapper[4777]: I1124 17:53:26.895173 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2d149ad6-8d86-445c-baed-c43f1eaec2cb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 17:53:26 crc kubenswrapper[4777]: I1124 17:53:26.895302 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b55r5\" (UniqueName: \"kubernetes.io/projected/2d149ad6-8d86-445c-baed-c43f1eaec2cb-kube-api-access-b55r5\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2d149ad6-8d86-445c-baed-c43f1eaec2cb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 17:53:26 crc kubenswrapper[4777]: I1124 17:53:26.895696 4777 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2d149ad6-8d86-445c-baed-c43f1eaec2cb\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 17:53:26 crc kubenswrapper[4777]: I1124 17:53:26.920094 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b55r5\" (UniqueName: \"kubernetes.io/projected/2d149ad6-8d86-445c-baed-c43f1eaec2cb-kube-api-access-b55r5\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2d149ad6-8d86-445c-baed-c43f1eaec2cb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 17:53:26 crc kubenswrapper[4777]: I1124 17:53:26.933187 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2d149ad6-8d86-445c-baed-c43f1eaec2cb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 17:53:27 crc kubenswrapper[4777]: I1124 17:53:27.050749 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 17:53:27 crc kubenswrapper[4777]: I1124 17:53:27.595136 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 17:53:27 crc kubenswrapper[4777]: I1124 17:53:27.601277 4777 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 17:53:27 crc kubenswrapper[4777]: I1124 17:53:27.862873 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"2d149ad6-8d86-445c-baed-c43f1eaec2cb","Type":"ContainerStarted","Data":"363e1108aa9e7f810e3ded9a7d0726d60fba0e89a8508601269e495b697f9971"} Nov 24 17:53:29 crc kubenswrapper[4777]: I1124 17:53:29.246174 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:53:29 crc kubenswrapper[4777]: E1124 17:53:29.247277 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:53:29 crc kubenswrapper[4777]: I1124 17:53:29.890924 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"2d149ad6-8d86-445c-baed-c43f1eaec2cb","Type":"ContainerStarted","Data":"7dcd3e1343b4e313935e4cdf6964f26c0a5e2b0bc768dfde7c7ea20f36fe2f3d"} Nov 24 17:53:29 crc kubenswrapper[4777]: I1124 17:53:29.913733 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.140200786 podStartE2EDuration="3.913706562s" podCreationTimestamp="2025-11-24 17:53:26 +0000 UTC" firstStartedPulling="2025-11-24 17:53:27.600831476 +0000 UTC m=+3165.760026565" lastFinishedPulling="2025-11-24 17:53:29.374337292 +0000 UTC m=+3167.533532341" observedRunningTime="2025-11-24 17:53:29.904413126 +0000 UTC m=+3168.063608195" watchObservedRunningTime="2025-11-24 17:53:29.913706562 +0000 UTC m=+3168.072901621" Nov 24 17:53:44 crc kubenswrapper[4777]: I1124 17:53:44.245091 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:53:44 crc kubenswrapper[4777]: E1124 17:53:44.246104 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:53:52 crc kubenswrapper[4777]: I1124 17:53:52.830092 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4gf5c/must-gather-krqqk"] Nov 24 17:53:52 crc kubenswrapper[4777]: I1124 17:53:52.832483 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/must-gather-krqqk" Nov 24 17:53:52 crc kubenswrapper[4777]: I1124 17:53:52.834497 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4gf5c"/"openshift-service-ca.crt" Nov 24 17:53:52 crc kubenswrapper[4777]: I1124 17:53:52.834775 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4gf5c"/"kube-root-ca.crt" Nov 24 17:53:52 crc kubenswrapper[4777]: I1124 17:53:52.855956 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4gf5c/must-gather-krqqk"] Nov 24 17:53:52 crc kubenswrapper[4777]: I1124 17:53:52.926614 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b575ae4b-aac8-4a05-a096-a85cae8f1718-must-gather-output\") pod \"must-gather-krqqk\" (UID: \"b575ae4b-aac8-4a05-a096-a85cae8f1718\") " pod="openshift-must-gather-4gf5c/must-gather-krqqk" Nov 24 17:53:52 crc kubenswrapper[4777]: I1124 17:53:52.926682 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cswr\" (UniqueName: \"kubernetes.io/projected/b575ae4b-aac8-4a05-a096-a85cae8f1718-kube-api-access-6cswr\") pod \"must-gather-krqqk\" (UID: \"b575ae4b-aac8-4a05-a096-a85cae8f1718\") " pod="openshift-must-gather-4gf5c/must-gather-krqqk" Nov 24 17:53:53 crc kubenswrapper[4777]: I1124 17:53:53.028599 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cswr\" (UniqueName: \"kubernetes.io/projected/b575ae4b-aac8-4a05-a096-a85cae8f1718-kube-api-access-6cswr\") pod \"must-gather-krqqk\" (UID: \"b575ae4b-aac8-4a05-a096-a85cae8f1718\") " pod="openshift-must-gather-4gf5c/must-gather-krqqk" Nov 24 17:53:53 crc kubenswrapper[4777]: I1124 17:53:53.028784 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b575ae4b-aac8-4a05-a096-a85cae8f1718-must-gather-output\") pod \"must-gather-krqqk\" (UID: \"b575ae4b-aac8-4a05-a096-a85cae8f1718\") " pod="openshift-must-gather-4gf5c/must-gather-krqqk" Nov 24 17:53:53 crc kubenswrapper[4777]: I1124 17:53:53.029259 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b575ae4b-aac8-4a05-a096-a85cae8f1718-must-gather-output\") pod \"must-gather-krqqk\" (UID: \"b575ae4b-aac8-4a05-a096-a85cae8f1718\") " pod="openshift-must-gather-4gf5c/must-gather-krqqk" Nov 24 17:53:53 crc kubenswrapper[4777]: I1124 17:53:53.066526 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cswr\" (UniqueName: \"kubernetes.io/projected/b575ae4b-aac8-4a05-a096-a85cae8f1718-kube-api-access-6cswr\") pod \"must-gather-krqqk\" (UID: \"b575ae4b-aac8-4a05-a096-a85cae8f1718\") " pod="openshift-must-gather-4gf5c/must-gather-krqqk" Nov 24 17:53:53 crc kubenswrapper[4777]: I1124 17:53:53.150927 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/must-gather-krqqk" Nov 24 17:53:53 crc kubenswrapper[4777]: I1124 17:53:53.721381 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4gf5c/must-gather-krqqk"] Nov 24 17:53:54 crc kubenswrapper[4777]: I1124 17:53:54.147987 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gf5c/must-gather-krqqk" event={"ID":"b575ae4b-aac8-4a05-a096-a85cae8f1718","Type":"ContainerStarted","Data":"d8861dc48e5a360d02c3b7090ab2d56cae26d37f541289143df215d36e2e4ecb"} Nov 24 17:53:56 crc kubenswrapper[4777]: I1124 17:53:56.245411 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:53:56 crc kubenswrapper[4777]: E1124 17:53:56.246018 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:54:00 crc kubenswrapper[4777]: I1124 17:54:00.277412 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gf5c/must-gather-krqqk" event={"ID":"b575ae4b-aac8-4a05-a096-a85cae8f1718","Type":"ContainerStarted","Data":"aa484d797efb71d3a21014ea068c0452dfd7ff17f9329a0b23417c7d9c675207"} Nov 24 17:54:00 crc kubenswrapper[4777]: I1124 17:54:00.277922 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gf5c/must-gather-krqqk" event={"ID":"b575ae4b-aac8-4a05-a096-a85cae8f1718","Type":"ContainerStarted","Data":"3b073f64a3f303b2e807761c1532e2ea625276750f3196db0dff67b43d5828d2"} Nov 24 17:54:00 crc kubenswrapper[4777]: I1124 17:54:00.311355 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4gf5c/must-gather-krqqk" podStartSLOduration=2.696462421 podStartE2EDuration="8.311337719s" podCreationTimestamp="2025-11-24 17:53:52 +0000 UTC" firstStartedPulling="2025-11-24 17:53:53.735613905 +0000 UTC m=+3191.894808954" lastFinishedPulling="2025-11-24 17:53:59.350489203 +0000 UTC m=+3197.509684252" observedRunningTime="2025-11-24 17:54:00.306613923 +0000 UTC m=+3198.465808972" watchObservedRunningTime="2025-11-24 17:54:00.311337719 +0000 UTC m=+3198.470532768" Nov 24 17:54:03 crc kubenswrapper[4777]: I1124 17:54:03.445763 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4gf5c/crc-debug-wvfwv"] Nov 24 17:54:03 crc kubenswrapper[4777]: I1124 17:54:03.447539 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/crc-debug-wvfwv" Nov 24 17:54:03 crc kubenswrapper[4777]: I1124 17:54:03.449843 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4gf5c"/"default-dockercfg-srvcj" Nov 24 17:54:03 crc kubenswrapper[4777]: I1124 17:54:03.576420 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4fc4b4af-ce10-47fa-80b7-95db7a0f2309-host\") pod \"crc-debug-wvfwv\" (UID: \"4fc4b4af-ce10-47fa-80b7-95db7a0f2309\") " pod="openshift-must-gather-4gf5c/crc-debug-wvfwv" Nov 24 17:54:03 crc kubenswrapper[4777]: I1124 17:54:03.576926 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6nqg\" (UniqueName: \"kubernetes.io/projected/4fc4b4af-ce10-47fa-80b7-95db7a0f2309-kube-api-access-b6nqg\") pod \"crc-debug-wvfwv\" (UID: \"4fc4b4af-ce10-47fa-80b7-95db7a0f2309\") " pod="openshift-must-gather-4gf5c/crc-debug-wvfwv" Nov 24 17:54:03 crc kubenswrapper[4777]: I1124 17:54:03.679513 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6nqg\" (UniqueName: \"kubernetes.io/projected/4fc4b4af-ce10-47fa-80b7-95db7a0f2309-kube-api-access-b6nqg\") pod \"crc-debug-wvfwv\" (UID: \"4fc4b4af-ce10-47fa-80b7-95db7a0f2309\") " pod="openshift-must-gather-4gf5c/crc-debug-wvfwv" Nov 24 17:54:03 crc kubenswrapper[4777]: I1124 17:54:03.679651 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4fc4b4af-ce10-47fa-80b7-95db7a0f2309-host\") pod \"crc-debug-wvfwv\" (UID: \"4fc4b4af-ce10-47fa-80b7-95db7a0f2309\") " pod="openshift-must-gather-4gf5c/crc-debug-wvfwv" Nov 24 17:54:03 crc kubenswrapper[4777]: I1124 17:54:03.679783 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4fc4b4af-ce10-47fa-80b7-95db7a0f2309-host\") pod \"crc-debug-wvfwv\" (UID: \"4fc4b4af-ce10-47fa-80b7-95db7a0f2309\") " pod="openshift-must-gather-4gf5c/crc-debug-wvfwv" Nov 24 17:54:03 crc kubenswrapper[4777]: I1124 17:54:03.709084 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6nqg\" (UniqueName: \"kubernetes.io/projected/4fc4b4af-ce10-47fa-80b7-95db7a0f2309-kube-api-access-b6nqg\") pod \"crc-debug-wvfwv\" (UID: \"4fc4b4af-ce10-47fa-80b7-95db7a0f2309\") " pod="openshift-must-gather-4gf5c/crc-debug-wvfwv" Nov 24 17:54:03 crc kubenswrapper[4777]: I1124 17:54:03.765858 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/crc-debug-wvfwv" Nov 24 17:54:03 crc kubenswrapper[4777]: W1124 17:54:03.801841 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fc4b4af_ce10_47fa_80b7_95db7a0f2309.slice/crio-026bda4591ba7fcbc8e63851c4172e1595c223f37e064da162d5236da3b5ca36 WatchSource:0}: Error finding container 026bda4591ba7fcbc8e63851c4172e1595c223f37e064da162d5236da3b5ca36: Status 404 returned error can't find the container with id 026bda4591ba7fcbc8e63851c4172e1595c223f37e064da162d5236da3b5ca36 Nov 24 17:54:04 crc kubenswrapper[4777]: I1124 17:54:04.319823 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gf5c/crc-debug-wvfwv" event={"ID":"4fc4b4af-ce10-47fa-80b7-95db7a0f2309","Type":"ContainerStarted","Data":"026bda4591ba7fcbc8e63851c4172e1595c223f37e064da162d5236da3b5ca36"} Nov 24 17:54:08 crc kubenswrapper[4777]: I1124 17:54:08.245535 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:54:08 crc kubenswrapper[4777]: E1124 17:54:08.246369 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 17:54:18 crc kubenswrapper[4777]: I1124 17:54:18.467590 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gf5c/crc-debug-wvfwv" event={"ID":"4fc4b4af-ce10-47fa-80b7-95db7a0f2309","Type":"ContainerStarted","Data":"3f2bd9573e432849a33721695ebce5aa6c74dc41eecd589d789da087cd988f90"} Nov 24 17:54:18 crc kubenswrapper[4777]: I1124 17:54:18.488255 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4gf5c/crc-debug-wvfwv" podStartSLOduration=1.431317023 podStartE2EDuration="15.488234219s" podCreationTimestamp="2025-11-24 17:54:03 +0000 UTC" firstStartedPulling="2025-11-24 17:54:03.804879558 +0000 UTC m=+3201.964074607" lastFinishedPulling="2025-11-24 17:54:17.861796754 +0000 UTC m=+3216.020991803" observedRunningTime="2025-11-24 17:54:18.480716564 +0000 UTC m=+3216.639911633" watchObservedRunningTime="2025-11-24 17:54:18.488234219 +0000 UTC m=+3216.647429268" Nov 24 17:54:23 crc kubenswrapper[4777]: I1124 17:54:23.252169 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:54:23 crc kubenswrapper[4777]: I1124 17:54:23.523355 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"4232c66077cf2691260b3a2ead0239df3749a3cf336096399a2eb7ed33863917"} Nov 24 17:55:09 crc kubenswrapper[4777]: I1124 17:55:09.998155 4777 generic.go:334] "Generic (PLEG): container finished" podID="4fc4b4af-ce10-47fa-80b7-95db7a0f2309" containerID="3f2bd9573e432849a33721695ebce5aa6c74dc41eecd589d789da087cd988f90" exitCode=0 Nov 24 17:55:09 crc kubenswrapper[4777]: I1124 17:55:09.998248 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gf5c/crc-debug-wvfwv" event={"ID":"4fc4b4af-ce10-47fa-80b7-95db7a0f2309","Type":"ContainerDied","Data":"3f2bd9573e432849a33721695ebce5aa6c74dc41eecd589d789da087cd988f90"} Nov 24 17:55:11 crc kubenswrapper[4777]: I1124 17:55:11.137158 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/crc-debug-wvfwv" Nov 24 17:55:11 crc kubenswrapper[4777]: I1124 17:55:11.176682 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4gf5c/crc-debug-wvfwv"] Nov 24 17:55:11 crc kubenswrapper[4777]: I1124 17:55:11.185556 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4gf5c/crc-debug-wvfwv"] Nov 24 17:55:11 crc kubenswrapper[4777]: I1124 17:55:11.281688 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6nqg\" (UniqueName: \"kubernetes.io/projected/4fc4b4af-ce10-47fa-80b7-95db7a0f2309-kube-api-access-b6nqg\") pod \"4fc4b4af-ce10-47fa-80b7-95db7a0f2309\" (UID: \"4fc4b4af-ce10-47fa-80b7-95db7a0f2309\") " Nov 24 17:55:11 crc kubenswrapper[4777]: I1124 17:55:11.281879 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4fc4b4af-ce10-47fa-80b7-95db7a0f2309-host\") pod \"4fc4b4af-ce10-47fa-80b7-95db7a0f2309\" (UID: \"4fc4b4af-ce10-47fa-80b7-95db7a0f2309\") " Nov 24 17:55:11 crc kubenswrapper[4777]: I1124 17:55:11.281930 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4fc4b4af-ce10-47fa-80b7-95db7a0f2309-host" (OuterVolumeSpecName: "host") pod "4fc4b4af-ce10-47fa-80b7-95db7a0f2309" (UID: "4fc4b4af-ce10-47fa-80b7-95db7a0f2309"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:55:11 crc kubenswrapper[4777]: I1124 17:55:11.285560 4777 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4fc4b4af-ce10-47fa-80b7-95db7a0f2309-host\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:11 crc kubenswrapper[4777]: I1124 17:55:11.287381 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fc4b4af-ce10-47fa-80b7-95db7a0f2309-kube-api-access-b6nqg" (OuterVolumeSpecName: "kube-api-access-b6nqg") pod "4fc4b4af-ce10-47fa-80b7-95db7a0f2309" (UID: "4fc4b4af-ce10-47fa-80b7-95db7a0f2309"). InnerVolumeSpecName "kube-api-access-b6nqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:55:11 crc kubenswrapper[4777]: I1124 17:55:11.389564 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6nqg\" (UniqueName: \"kubernetes.io/projected/4fc4b4af-ce10-47fa-80b7-95db7a0f2309-kube-api-access-b6nqg\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:12 crc kubenswrapper[4777]: I1124 17:55:12.020708 4777 scope.go:117] "RemoveContainer" containerID="3f2bd9573e432849a33721695ebce5aa6c74dc41eecd589d789da087cd988f90" Nov 24 17:55:12 crc kubenswrapper[4777]: I1124 17:55:12.020795 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/crc-debug-wvfwv" Nov 24 17:55:12 crc kubenswrapper[4777]: I1124 17:55:12.348826 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4gf5c/crc-debug-xl269"] Nov 24 17:55:12 crc kubenswrapper[4777]: E1124 17:55:12.349276 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fc4b4af-ce10-47fa-80b7-95db7a0f2309" containerName="container-00" Nov 24 17:55:12 crc kubenswrapper[4777]: I1124 17:55:12.349290 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fc4b4af-ce10-47fa-80b7-95db7a0f2309" containerName="container-00" Nov 24 17:55:12 crc kubenswrapper[4777]: I1124 17:55:12.349498 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fc4b4af-ce10-47fa-80b7-95db7a0f2309" containerName="container-00" Nov 24 17:55:12 crc kubenswrapper[4777]: I1124 17:55:12.350439 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/crc-debug-xl269" Nov 24 17:55:12 crc kubenswrapper[4777]: I1124 17:55:12.353257 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4gf5c"/"default-dockercfg-srvcj" Nov 24 17:55:12 crc kubenswrapper[4777]: I1124 17:55:12.410336 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdkhd\" (UniqueName: \"kubernetes.io/projected/b806f3ac-dd20-457b-8bbc-5a5490969067-kube-api-access-mdkhd\") pod \"crc-debug-xl269\" (UID: \"b806f3ac-dd20-457b-8bbc-5a5490969067\") " pod="openshift-must-gather-4gf5c/crc-debug-xl269" Nov 24 17:55:12 crc kubenswrapper[4777]: I1124 17:55:12.410453 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b806f3ac-dd20-457b-8bbc-5a5490969067-host\") pod \"crc-debug-xl269\" (UID: \"b806f3ac-dd20-457b-8bbc-5a5490969067\") " pod="openshift-must-gather-4gf5c/crc-debug-xl269" Nov 24 17:55:12 crc kubenswrapper[4777]: I1124 17:55:12.512690 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdkhd\" (UniqueName: \"kubernetes.io/projected/b806f3ac-dd20-457b-8bbc-5a5490969067-kube-api-access-mdkhd\") pod \"crc-debug-xl269\" (UID: \"b806f3ac-dd20-457b-8bbc-5a5490969067\") " pod="openshift-must-gather-4gf5c/crc-debug-xl269" Nov 24 17:55:12 crc kubenswrapper[4777]: I1124 17:55:12.513149 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b806f3ac-dd20-457b-8bbc-5a5490969067-host\") pod \"crc-debug-xl269\" (UID: \"b806f3ac-dd20-457b-8bbc-5a5490969067\") " pod="openshift-must-gather-4gf5c/crc-debug-xl269" Nov 24 17:55:12 crc kubenswrapper[4777]: I1124 17:55:12.513292 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b806f3ac-dd20-457b-8bbc-5a5490969067-host\") pod \"crc-debug-xl269\" (UID: \"b806f3ac-dd20-457b-8bbc-5a5490969067\") " pod="openshift-must-gather-4gf5c/crc-debug-xl269" Nov 24 17:55:12 crc kubenswrapper[4777]: I1124 17:55:12.553135 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdkhd\" (UniqueName: \"kubernetes.io/projected/b806f3ac-dd20-457b-8bbc-5a5490969067-kube-api-access-mdkhd\") pod \"crc-debug-xl269\" (UID: \"b806f3ac-dd20-457b-8bbc-5a5490969067\") " pod="openshift-must-gather-4gf5c/crc-debug-xl269" Nov 24 17:55:12 crc kubenswrapper[4777]: I1124 17:55:12.677126 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/crc-debug-xl269" Nov 24 17:55:13 crc kubenswrapper[4777]: I1124 17:55:13.033881 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gf5c/crc-debug-xl269" event={"ID":"b806f3ac-dd20-457b-8bbc-5a5490969067","Type":"ContainerStarted","Data":"e0203afc708a619716fcce5fafa3d9fe7cb926c114f86b2f7e40125c3b3e18b0"} Nov 24 17:55:13 crc kubenswrapper[4777]: I1124 17:55:13.034279 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gf5c/crc-debug-xl269" event={"ID":"b806f3ac-dd20-457b-8bbc-5a5490969067","Type":"ContainerStarted","Data":"ccb2e528393985afbc4bcfb4b33ae29d871d0d40a0532f54be13490229d1f4a8"} Nov 24 17:55:13 crc kubenswrapper[4777]: I1124 17:55:13.054533 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4gf5c/crc-debug-xl269" podStartSLOduration=1.054512063 podStartE2EDuration="1.054512063s" podCreationTimestamp="2025-11-24 17:55:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 17:55:13.043336853 +0000 UTC m=+3271.202531912" watchObservedRunningTime="2025-11-24 17:55:13.054512063 +0000 UTC m=+3271.213707132" Nov 24 17:55:13 crc kubenswrapper[4777]: I1124 17:55:13.254939 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fc4b4af-ce10-47fa-80b7-95db7a0f2309" path="/var/lib/kubelet/pods/4fc4b4af-ce10-47fa-80b7-95db7a0f2309/volumes" Nov 24 17:55:14 crc kubenswrapper[4777]: I1124 17:55:14.052833 4777 generic.go:334] "Generic (PLEG): container finished" podID="b806f3ac-dd20-457b-8bbc-5a5490969067" containerID="e0203afc708a619716fcce5fafa3d9fe7cb926c114f86b2f7e40125c3b3e18b0" exitCode=0 Nov 24 17:55:14 crc kubenswrapper[4777]: I1124 17:55:14.053191 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gf5c/crc-debug-xl269" event={"ID":"b806f3ac-dd20-457b-8bbc-5a5490969067","Type":"ContainerDied","Data":"e0203afc708a619716fcce5fafa3d9fe7cb926c114f86b2f7e40125c3b3e18b0"} Nov 24 17:55:15 crc kubenswrapper[4777]: I1124 17:55:15.194026 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/crc-debug-xl269" Nov 24 17:55:15 crc kubenswrapper[4777]: I1124 17:55:15.238401 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4gf5c/crc-debug-xl269"] Nov 24 17:55:15 crc kubenswrapper[4777]: I1124 17:55:15.265462 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4gf5c/crc-debug-xl269"] Nov 24 17:55:15 crc kubenswrapper[4777]: I1124 17:55:15.364001 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdkhd\" (UniqueName: \"kubernetes.io/projected/b806f3ac-dd20-457b-8bbc-5a5490969067-kube-api-access-mdkhd\") pod \"b806f3ac-dd20-457b-8bbc-5a5490969067\" (UID: \"b806f3ac-dd20-457b-8bbc-5a5490969067\") " Nov 24 17:55:15 crc kubenswrapper[4777]: I1124 17:55:15.364250 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b806f3ac-dd20-457b-8bbc-5a5490969067-host\") pod \"b806f3ac-dd20-457b-8bbc-5a5490969067\" (UID: \"b806f3ac-dd20-457b-8bbc-5a5490969067\") " Nov 24 17:55:15 crc kubenswrapper[4777]: I1124 17:55:15.364322 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b806f3ac-dd20-457b-8bbc-5a5490969067-host" (OuterVolumeSpecName: "host") pod "b806f3ac-dd20-457b-8bbc-5a5490969067" (UID: "b806f3ac-dd20-457b-8bbc-5a5490969067"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:55:15 crc kubenswrapper[4777]: I1124 17:55:15.364907 4777 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b806f3ac-dd20-457b-8bbc-5a5490969067-host\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:15 crc kubenswrapper[4777]: I1124 17:55:15.371625 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b806f3ac-dd20-457b-8bbc-5a5490969067-kube-api-access-mdkhd" (OuterVolumeSpecName: "kube-api-access-mdkhd") pod "b806f3ac-dd20-457b-8bbc-5a5490969067" (UID: "b806f3ac-dd20-457b-8bbc-5a5490969067"). InnerVolumeSpecName "kube-api-access-mdkhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:55:15 crc kubenswrapper[4777]: I1124 17:55:15.466514 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdkhd\" (UniqueName: \"kubernetes.io/projected/b806f3ac-dd20-457b-8bbc-5a5490969067-kube-api-access-mdkhd\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:16 crc kubenswrapper[4777]: I1124 17:55:16.079554 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccb2e528393985afbc4bcfb4b33ae29d871d0d40a0532f54be13490229d1f4a8" Nov 24 17:55:16 crc kubenswrapper[4777]: I1124 17:55:16.079619 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/crc-debug-xl269" Nov 24 17:55:16 crc kubenswrapper[4777]: I1124 17:55:16.551717 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4gf5c/crc-debug-rpsjf"] Nov 24 17:55:16 crc kubenswrapper[4777]: E1124 17:55:16.552584 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b806f3ac-dd20-457b-8bbc-5a5490969067" containerName="container-00" Nov 24 17:55:16 crc kubenswrapper[4777]: I1124 17:55:16.552600 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="b806f3ac-dd20-457b-8bbc-5a5490969067" containerName="container-00" Nov 24 17:55:16 crc kubenswrapper[4777]: I1124 17:55:16.552849 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="b806f3ac-dd20-457b-8bbc-5a5490969067" containerName="container-00" Nov 24 17:55:16 crc kubenswrapper[4777]: I1124 17:55:16.553713 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/crc-debug-rpsjf" Nov 24 17:55:16 crc kubenswrapper[4777]: I1124 17:55:16.555642 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4gf5c"/"default-dockercfg-srvcj" Nov 24 17:55:16 crc kubenswrapper[4777]: I1124 17:55:16.688032 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9cdf32e-6a91-406b-b5aa-77eb65a926d5-host\") pod \"crc-debug-rpsjf\" (UID: \"f9cdf32e-6a91-406b-b5aa-77eb65a926d5\") " pod="openshift-must-gather-4gf5c/crc-debug-rpsjf" Nov 24 17:55:16 crc kubenswrapper[4777]: I1124 17:55:16.688191 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvgsq\" (UniqueName: \"kubernetes.io/projected/f9cdf32e-6a91-406b-b5aa-77eb65a926d5-kube-api-access-hvgsq\") pod \"crc-debug-rpsjf\" (UID: \"f9cdf32e-6a91-406b-b5aa-77eb65a926d5\") " pod="openshift-must-gather-4gf5c/crc-debug-rpsjf" Nov 24 17:55:16 crc kubenswrapper[4777]: I1124 17:55:16.816929 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9cdf32e-6a91-406b-b5aa-77eb65a926d5-host\") pod \"crc-debug-rpsjf\" (UID: \"f9cdf32e-6a91-406b-b5aa-77eb65a926d5\") " pod="openshift-must-gather-4gf5c/crc-debug-rpsjf" Nov 24 17:55:16 crc kubenswrapper[4777]: I1124 17:55:16.817327 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvgsq\" (UniqueName: \"kubernetes.io/projected/f9cdf32e-6a91-406b-b5aa-77eb65a926d5-kube-api-access-hvgsq\") pod \"crc-debug-rpsjf\" (UID: \"f9cdf32e-6a91-406b-b5aa-77eb65a926d5\") " pod="openshift-must-gather-4gf5c/crc-debug-rpsjf" Nov 24 17:55:16 crc kubenswrapper[4777]: I1124 17:55:16.818037 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9cdf32e-6a91-406b-b5aa-77eb65a926d5-host\") pod \"crc-debug-rpsjf\" (UID: \"f9cdf32e-6a91-406b-b5aa-77eb65a926d5\") " pod="openshift-must-gather-4gf5c/crc-debug-rpsjf" Nov 24 17:55:16 crc kubenswrapper[4777]: I1124 17:55:16.879828 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvgsq\" (UniqueName: \"kubernetes.io/projected/f9cdf32e-6a91-406b-b5aa-77eb65a926d5-kube-api-access-hvgsq\") pod \"crc-debug-rpsjf\" (UID: \"f9cdf32e-6a91-406b-b5aa-77eb65a926d5\") " pod="openshift-must-gather-4gf5c/crc-debug-rpsjf" Nov 24 17:55:17 crc kubenswrapper[4777]: I1124 17:55:17.170370 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/crc-debug-rpsjf" Nov 24 17:55:17 crc kubenswrapper[4777]: W1124 17:55:17.206201 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9cdf32e_6a91_406b_b5aa_77eb65a926d5.slice/crio-f6ca54f32e59039e296c7db599e5456a27a8963f24ee10725e656ebe8feeedc0 WatchSource:0}: Error finding container f6ca54f32e59039e296c7db599e5456a27a8963f24ee10725e656ebe8feeedc0: Status 404 returned error can't find the container with id f6ca54f32e59039e296c7db599e5456a27a8963f24ee10725e656ebe8feeedc0 Nov 24 17:55:17 crc kubenswrapper[4777]: I1124 17:55:17.266943 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b806f3ac-dd20-457b-8bbc-5a5490969067" path="/var/lib/kubelet/pods/b806f3ac-dd20-457b-8bbc-5a5490969067/volumes" Nov 24 17:55:18 crc kubenswrapper[4777]: I1124 17:55:18.099033 4777 generic.go:334] "Generic (PLEG): container finished" podID="f9cdf32e-6a91-406b-b5aa-77eb65a926d5" containerID="e1f0aabba2e780713c4954a1f1907720f889b32c2ee3f711254168dedec6b37a" exitCode=0 Nov 24 17:55:18 crc kubenswrapper[4777]: I1124 17:55:18.099102 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gf5c/crc-debug-rpsjf" event={"ID":"f9cdf32e-6a91-406b-b5aa-77eb65a926d5","Type":"ContainerDied","Data":"e1f0aabba2e780713c4954a1f1907720f889b32c2ee3f711254168dedec6b37a"} Nov 24 17:55:18 crc kubenswrapper[4777]: I1124 17:55:18.099150 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gf5c/crc-debug-rpsjf" event={"ID":"f9cdf32e-6a91-406b-b5aa-77eb65a926d5","Type":"ContainerStarted","Data":"f6ca54f32e59039e296c7db599e5456a27a8963f24ee10725e656ebe8feeedc0"} Nov 24 17:55:18 crc kubenswrapper[4777]: I1124 17:55:18.133597 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4gf5c/crc-debug-rpsjf"] Nov 24 17:55:18 crc kubenswrapper[4777]: I1124 17:55:18.143142 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4gf5c/crc-debug-rpsjf"] Nov 24 17:55:19 crc kubenswrapper[4777]: I1124 17:55:19.236663 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/crc-debug-rpsjf" Nov 24 17:55:19 crc kubenswrapper[4777]: I1124 17:55:19.376570 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvgsq\" (UniqueName: \"kubernetes.io/projected/f9cdf32e-6a91-406b-b5aa-77eb65a926d5-kube-api-access-hvgsq\") pod \"f9cdf32e-6a91-406b-b5aa-77eb65a926d5\" (UID: \"f9cdf32e-6a91-406b-b5aa-77eb65a926d5\") " Nov 24 17:55:19 crc kubenswrapper[4777]: I1124 17:55:19.376741 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9cdf32e-6a91-406b-b5aa-77eb65a926d5-host\") pod \"f9cdf32e-6a91-406b-b5aa-77eb65a926d5\" (UID: \"f9cdf32e-6a91-406b-b5aa-77eb65a926d5\") " Nov 24 17:55:19 crc kubenswrapper[4777]: I1124 17:55:19.378629 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9cdf32e-6a91-406b-b5aa-77eb65a926d5-host" (OuterVolumeSpecName: "host") pod "f9cdf32e-6a91-406b-b5aa-77eb65a926d5" (UID: "f9cdf32e-6a91-406b-b5aa-77eb65a926d5"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 17:55:19 crc kubenswrapper[4777]: I1124 17:55:19.391628 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9cdf32e-6a91-406b-b5aa-77eb65a926d5-kube-api-access-hvgsq" (OuterVolumeSpecName: "kube-api-access-hvgsq") pod "f9cdf32e-6a91-406b-b5aa-77eb65a926d5" (UID: "f9cdf32e-6a91-406b-b5aa-77eb65a926d5"). InnerVolumeSpecName "kube-api-access-hvgsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:55:19 crc kubenswrapper[4777]: I1124 17:55:19.479495 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvgsq\" (UniqueName: \"kubernetes.io/projected/f9cdf32e-6a91-406b-b5aa-77eb65a926d5-kube-api-access-hvgsq\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:19 crc kubenswrapper[4777]: I1124 17:55:19.479528 4777 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9cdf32e-6a91-406b-b5aa-77eb65a926d5-host\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:20 crc kubenswrapper[4777]: I1124 17:55:20.117213 4777 scope.go:117] "RemoveContainer" containerID="e1f0aabba2e780713c4954a1f1907720f889b32c2ee3f711254168dedec6b37a" Nov 24 17:55:20 crc kubenswrapper[4777]: I1124 17:55:20.117263 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/crc-debug-rpsjf" Nov 24 17:55:21 crc kubenswrapper[4777]: I1124 17:55:21.280366 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9cdf32e-6a91-406b-b5aa-77eb65a926d5" path="/var/lib/kubelet/pods/f9cdf32e-6a91-406b-b5aa-77eb65a926d5/volumes" Nov 24 17:55:26 crc kubenswrapper[4777]: I1124 17:55:26.952525 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6jtbv"] Nov 24 17:55:26 crc kubenswrapper[4777]: E1124 17:55:26.953449 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9cdf32e-6a91-406b-b5aa-77eb65a926d5" containerName="container-00" Nov 24 17:55:26 crc kubenswrapper[4777]: I1124 17:55:26.953463 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9cdf32e-6a91-406b-b5aa-77eb65a926d5" containerName="container-00" Nov 24 17:55:26 crc kubenswrapper[4777]: I1124 17:55:26.953741 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9cdf32e-6a91-406b-b5aa-77eb65a926d5" containerName="container-00" Nov 24 17:55:26 crc kubenswrapper[4777]: I1124 17:55:26.955387 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:26 crc kubenswrapper[4777]: I1124 17:55:26.977401 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6jtbv"] Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.086076 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xczx6\" (UniqueName: \"kubernetes.io/projected/91f65ef1-339d-486f-8fe1-4670be8a14e7-kube-api-access-xczx6\") pod \"certified-operators-6jtbv\" (UID: \"91f65ef1-339d-486f-8fe1-4670be8a14e7\") " pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.086419 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f65ef1-339d-486f-8fe1-4670be8a14e7-utilities\") pod \"certified-operators-6jtbv\" (UID: \"91f65ef1-339d-486f-8fe1-4670be8a14e7\") " pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.086445 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f65ef1-339d-486f-8fe1-4670be8a14e7-catalog-content\") pod \"certified-operators-6jtbv\" (UID: \"91f65ef1-339d-486f-8fe1-4670be8a14e7\") " pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.152544 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fm4mt"] Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.154728 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.177823 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fm4mt"] Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.189864 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xczx6\" (UniqueName: \"kubernetes.io/projected/91f65ef1-339d-486f-8fe1-4670be8a14e7-kube-api-access-xczx6\") pod \"certified-operators-6jtbv\" (UID: \"91f65ef1-339d-486f-8fe1-4670be8a14e7\") " pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.189980 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f65ef1-339d-486f-8fe1-4670be8a14e7-utilities\") pod \"certified-operators-6jtbv\" (UID: \"91f65ef1-339d-486f-8fe1-4670be8a14e7\") " pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.190017 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f65ef1-339d-486f-8fe1-4670be8a14e7-catalog-content\") pod \"certified-operators-6jtbv\" (UID: \"91f65ef1-339d-486f-8fe1-4670be8a14e7\") " pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.190619 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f65ef1-339d-486f-8fe1-4670be8a14e7-catalog-content\") pod \"certified-operators-6jtbv\" (UID: \"91f65ef1-339d-486f-8fe1-4670be8a14e7\") " pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.192090 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f65ef1-339d-486f-8fe1-4670be8a14e7-utilities\") pod \"certified-operators-6jtbv\" (UID: \"91f65ef1-339d-486f-8fe1-4670be8a14e7\") " pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.235190 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xczx6\" (UniqueName: \"kubernetes.io/projected/91f65ef1-339d-486f-8fe1-4670be8a14e7-kube-api-access-xczx6\") pod \"certified-operators-6jtbv\" (UID: \"91f65ef1-339d-486f-8fe1-4670be8a14e7\") " pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.290502 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.291760 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e418e82-f285-4adf-a6ac-e425aa66c9c8-catalog-content\") pod \"redhat-operators-fm4mt\" (UID: \"7e418e82-f285-4adf-a6ac-e425aa66c9c8\") " pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.291868 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4646r\" (UniqueName: \"kubernetes.io/projected/7e418e82-f285-4adf-a6ac-e425aa66c9c8-kube-api-access-4646r\") pod \"redhat-operators-fm4mt\" (UID: \"7e418e82-f285-4adf-a6ac-e425aa66c9c8\") " pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.292121 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e418e82-f285-4adf-a6ac-e425aa66c9c8-utilities\") pod \"redhat-operators-fm4mt\" (UID: \"7e418e82-f285-4adf-a6ac-e425aa66c9c8\") " pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.393991 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e418e82-f285-4adf-a6ac-e425aa66c9c8-catalog-content\") pod \"redhat-operators-fm4mt\" (UID: \"7e418e82-f285-4adf-a6ac-e425aa66c9c8\") " pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.394054 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4646r\" (UniqueName: \"kubernetes.io/projected/7e418e82-f285-4adf-a6ac-e425aa66c9c8-kube-api-access-4646r\") pod \"redhat-operators-fm4mt\" (UID: \"7e418e82-f285-4adf-a6ac-e425aa66c9c8\") " pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.394167 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e418e82-f285-4adf-a6ac-e425aa66c9c8-utilities\") pod \"redhat-operators-fm4mt\" (UID: \"7e418e82-f285-4adf-a6ac-e425aa66c9c8\") " pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.394865 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e418e82-f285-4adf-a6ac-e425aa66c9c8-utilities\") pod \"redhat-operators-fm4mt\" (UID: \"7e418e82-f285-4adf-a6ac-e425aa66c9c8\") " pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.395153 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e418e82-f285-4adf-a6ac-e425aa66c9c8-catalog-content\") pod \"redhat-operators-fm4mt\" (UID: \"7e418e82-f285-4adf-a6ac-e425aa66c9c8\") " pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.415714 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4646r\" (UniqueName: \"kubernetes.io/projected/7e418e82-f285-4adf-a6ac-e425aa66c9c8-kube-api-access-4646r\") pod \"redhat-operators-fm4mt\" (UID: \"7e418e82-f285-4adf-a6ac-e425aa66c9c8\") " pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.489652 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:55:27 crc kubenswrapper[4777]: I1124 17:55:27.944028 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6jtbv"] Nov 24 17:55:28 crc kubenswrapper[4777]: I1124 17:55:28.208026 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jtbv" event={"ID":"91f65ef1-339d-486f-8fe1-4670be8a14e7","Type":"ContainerStarted","Data":"28bd6ce70d4653e493f82d1f39d6f8e6d473f694e0eaf8706f80f8208306ac76"} Nov 24 17:55:28 crc kubenswrapper[4777]: W1124 17:55:28.232755 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e418e82_f285_4adf_a6ac_e425aa66c9c8.slice/crio-6a0e1b5eb4ccda19a255d236d5a900238fca67d56560f5f15f91398b0b840388 WatchSource:0}: Error finding container 6a0e1b5eb4ccda19a255d236d5a900238fca67d56560f5f15f91398b0b840388: Status 404 returned error can't find the container with id 6a0e1b5eb4ccda19a255d236d5a900238fca67d56560f5f15f91398b0b840388 Nov 24 17:55:28 crc kubenswrapper[4777]: I1124 17:55:28.235301 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fm4mt"] Nov 24 17:55:29 crc kubenswrapper[4777]: I1124 17:55:29.217350 4777 generic.go:334] "Generic (PLEG): container finished" podID="7e418e82-f285-4adf-a6ac-e425aa66c9c8" containerID="c180bbeef8ff04fde81dd992ed00d37aab8f5778c2a9750bd97c9323d0c76717" exitCode=0 Nov 24 17:55:29 crc kubenswrapper[4777]: I1124 17:55:29.217444 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fm4mt" event={"ID":"7e418e82-f285-4adf-a6ac-e425aa66c9c8","Type":"ContainerDied","Data":"c180bbeef8ff04fde81dd992ed00d37aab8f5778c2a9750bd97c9323d0c76717"} Nov 24 17:55:29 crc kubenswrapper[4777]: I1124 17:55:29.217951 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fm4mt" event={"ID":"7e418e82-f285-4adf-a6ac-e425aa66c9c8","Type":"ContainerStarted","Data":"6a0e1b5eb4ccda19a255d236d5a900238fca67d56560f5f15f91398b0b840388"} Nov 24 17:55:29 crc kubenswrapper[4777]: I1124 17:55:29.219776 4777 generic.go:334] "Generic (PLEG): container finished" podID="91f65ef1-339d-486f-8fe1-4670be8a14e7" containerID="4b9d047d2dc2d3149912fbf4ec2c6f90ed23af4973c4f1109e1702b15e12b4bb" exitCode=0 Nov 24 17:55:29 crc kubenswrapper[4777]: I1124 17:55:29.219819 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jtbv" event={"ID":"91f65ef1-339d-486f-8fe1-4670be8a14e7","Type":"ContainerDied","Data":"4b9d047d2dc2d3149912fbf4ec2c6f90ed23af4973c4f1109e1702b15e12b4bb"} Nov 24 17:55:30 crc kubenswrapper[4777]: I1124 17:55:30.253307 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fm4mt" event={"ID":"7e418e82-f285-4adf-a6ac-e425aa66c9c8","Type":"ContainerStarted","Data":"c3232a1911fb7751da9aeb7ccdd79d724709b42ac0d9a6b99b3a071c69a3b4a2"} Nov 24 17:55:30 crc kubenswrapper[4777]: I1124 17:55:30.256917 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jtbv" event={"ID":"91f65ef1-339d-486f-8fe1-4670be8a14e7","Type":"ContainerStarted","Data":"09d712a531e3554fb4bb18b90ce8fb2fa4bf86ab9638118b719ffd0327b147f3"} Nov 24 17:55:33 crc kubenswrapper[4777]: I1124 17:55:33.286581 4777 generic.go:334] "Generic (PLEG): container finished" podID="91f65ef1-339d-486f-8fe1-4670be8a14e7" containerID="09d712a531e3554fb4bb18b90ce8fb2fa4bf86ab9638118b719ffd0327b147f3" exitCode=0 Nov 24 17:55:33 crc kubenswrapper[4777]: I1124 17:55:33.286630 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jtbv" event={"ID":"91f65ef1-339d-486f-8fe1-4670be8a14e7","Type":"ContainerDied","Data":"09d712a531e3554fb4bb18b90ce8fb2fa4bf86ab9638118b719ffd0327b147f3"} Nov 24 17:55:34 crc kubenswrapper[4777]: I1124 17:55:34.303018 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jtbv" event={"ID":"91f65ef1-339d-486f-8fe1-4670be8a14e7","Type":"ContainerStarted","Data":"c0845bf314c42044993e6aa74a3419bed10caffb54bd00ac54c380a45cd50582"} Nov 24 17:55:34 crc kubenswrapper[4777]: I1124 17:55:34.331463 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6jtbv" podStartSLOduration=3.800683418 podStartE2EDuration="8.331443721s" podCreationTimestamp="2025-11-24 17:55:26 +0000 UTC" firstStartedPulling="2025-11-24 17:55:29.222723551 +0000 UTC m=+3287.381918600" lastFinishedPulling="2025-11-24 17:55:33.753483854 +0000 UTC m=+3291.912678903" observedRunningTime="2025-11-24 17:55:34.32371516 +0000 UTC m=+3292.482910219" watchObservedRunningTime="2025-11-24 17:55:34.331443721 +0000 UTC m=+3292.490638780" Nov 24 17:55:37 crc kubenswrapper[4777]: I1124 17:55:37.291526 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:37 crc kubenswrapper[4777]: I1124 17:55:37.306898 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:37 crc kubenswrapper[4777]: I1124 17:55:37.339692 4777 generic.go:334] "Generic (PLEG): container finished" podID="7e418e82-f285-4adf-a6ac-e425aa66c9c8" containerID="c3232a1911fb7751da9aeb7ccdd79d724709b42ac0d9a6b99b3a071c69a3b4a2" exitCode=0 Nov 24 17:55:37 crc kubenswrapper[4777]: I1124 17:55:37.339764 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fm4mt" event={"ID":"7e418e82-f285-4adf-a6ac-e425aa66c9c8","Type":"ContainerDied","Data":"c3232a1911fb7751da9aeb7ccdd79d724709b42ac0d9a6b99b3a071c69a3b4a2"} Nov 24 17:55:38 crc kubenswrapper[4777]: I1124 17:55:38.339275 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-6jtbv" podUID="91f65ef1-339d-486f-8fe1-4670be8a14e7" containerName="registry-server" probeResult="failure" output=< Nov 24 17:55:38 crc kubenswrapper[4777]: timeout: failed to connect service ":50051" within 1s Nov 24 17:55:38 crc kubenswrapper[4777]: > Nov 24 17:55:38 crc kubenswrapper[4777]: I1124 17:55:38.350619 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fm4mt" event={"ID":"7e418e82-f285-4adf-a6ac-e425aa66c9c8","Type":"ContainerStarted","Data":"7a9fec5b3da6967d367fbf946d5ccc54ab9fc21561ae7675a4fbc5ced5d7d426"} Nov 24 17:55:38 crc kubenswrapper[4777]: I1124 17:55:38.376694 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fm4mt" podStartSLOduration=2.8456061249999998 podStartE2EDuration="11.376670064s" podCreationTimestamp="2025-11-24 17:55:27 +0000 UTC" firstStartedPulling="2025-11-24 17:55:29.220174618 +0000 UTC m=+3287.379369667" lastFinishedPulling="2025-11-24 17:55:37.751238557 +0000 UTC m=+3295.910433606" observedRunningTime="2025-11-24 17:55:38.368013866 +0000 UTC m=+3296.527208925" watchObservedRunningTime="2025-11-24 17:55:38.376670064 +0000 UTC m=+3296.535865113" Nov 24 17:55:46 crc kubenswrapper[4777]: I1124 17:55:46.878871 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_055503f1-ab1c-47db-be36-37a3f572be32/init-config-reloader/0.log" Nov 24 17:55:47 crc kubenswrapper[4777]: I1124 17:55:47.187616 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_055503f1-ab1c-47db-be36-37a3f572be32/init-config-reloader/0.log" Nov 24 17:55:47 crc kubenswrapper[4777]: I1124 17:55:47.308521 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_055503f1-ab1c-47db-be36-37a3f572be32/config-reloader/0.log" Nov 24 17:55:47 crc kubenswrapper[4777]: I1124 17:55:47.340598 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_055503f1-ab1c-47db-be36-37a3f572be32/alertmanager/0.log" Nov 24 17:55:47 crc kubenswrapper[4777]: I1124 17:55:47.365956 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:47 crc kubenswrapper[4777]: I1124 17:55:47.416677 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:47 crc kubenswrapper[4777]: I1124 17:55:47.489908 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:55:47 crc kubenswrapper[4777]: I1124 17:55:47.491520 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:55:47 crc kubenswrapper[4777]: I1124 17:55:47.507077 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-65bb697bdb-4pwlb_27a38d66-1471-470a-95e9-c4bfa5ba415b/barbican-api/0.log" Nov 24 17:55:47 crc kubenswrapper[4777]: I1124 17:55:47.605033 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6jtbv"] Nov 24 17:55:47 crc kubenswrapper[4777]: I1124 17:55:47.743745 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-65bb697bdb-4pwlb_27a38d66-1471-470a-95e9-c4bfa5ba415b/barbican-api-log/0.log" Nov 24 17:55:47 crc kubenswrapper[4777]: I1124 17:55:47.928713 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-845875d46-n8hgg_d173b040-b120-4c77-ba3c-b13c41a6ac37/barbican-keystone-listener/0.log" Nov 24 17:55:48 crc kubenswrapper[4777]: I1124 17:55:48.101082 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-845875d46-n8hgg_d173b040-b120-4c77-ba3c-b13c41a6ac37/barbican-keystone-listener-log/0.log" Nov 24 17:55:48 crc kubenswrapper[4777]: I1124 17:55:48.182739 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5d57df9989-fbq7p_dded6598-74bf-4395-88e0-37152b4dd16c/barbican-worker/0.log" Nov 24 17:55:48 crc kubenswrapper[4777]: I1124 17:55:48.203543 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5d57df9989-fbq7p_dded6598-74bf-4395-88e0-37152b4dd16c/barbican-worker-log/0.log" Nov 24 17:55:48 crc kubenswrapper[4777]: I1124 17:55:48.422794 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g_1bda09cc-b84e-409f-b4cd-fc387aa02c61/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:55:48 crc kubenswrapper[4777]: I1124 17:55:48.496915 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6jtbv" podUID="91f65ef1-339d-486f-8fe1-4670be8a14e7" containerName="registry-server" containerID="cri-o://c0845bf314c42044993e6aa74a3419bed10caffb54bd00ac54c380a45cd50582" gracePeriod=2 Nov 24 17:55:48 crc kubenswrapper[4777]: I1124 17:55:48.545533 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f80e34e-82d8-4d61-91b9-323898efdabf/ceilometer-central-agent/0.log" Nov 24 17:55:48 crc kubenswrapper[4777]: I1124 17:55:48.548680 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fm4mt" podUID="7e418e82-f285-4adf-a6ac-e425aa66c9c8" containerName="registry-server" probeResult="failure" output=< Nov 24 17:55:48 crc kubenswrapper[4777]: timeout: failed to connect service ":50051" within 1s Nov 24 17:55:48 crc kubenswrapper[4777]: > Nov 24 17:55:48 crc kubenswrapper[4777]: I1124 17:55:48.614718 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f80e34e-82d8-4d61-91b9-323898efdabf/ceilometer-notification-agent/0.log" Nov 24 17:55:48 crc kubenswrapper[4777]: I1124 17:55:48.696031 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f80e34e-82d8-4d61-91b9-323898efdabf/proxy-httpd/0.log" Nov 24 17:55:48 crc kubenswrapper[4777]: I1124 17:55:48.768061 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f80e34e-82d8-4d61-91b9-323898efdabf/sg-core/0.log" Nov 24 17:55:48 crc kubenswrapper[4777]: I1124 17:55:48.920558 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_20f43f58-1756-48e1-a958-b49bc82ef1b4/cinder-api/0.log" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.039777 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_20f43f58-1756-48e1-a958-b49bc82ef1b4/cinder-api-log/0.log" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.197947 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13/cinder-scheduler/0.log" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.224042 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13/probe/0.log" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.269647 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.370582 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f65ef1-339d-486f-8fe1-4670be8a14e7-catalog-content\") pod \"91f65ef1-339d-486f-8fe1-4670be8a14e7\" (UID: \"91f65ef1-339d-486f-8fe1-4670be8a14e7\") " Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.371166 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xczx6\" (UniqueName: \"kubernetes.io/projected/91f65ef1-339d-486f-8fe1-4670be8a14e7-kube-api-access-xczx6\") pod \"91f65ef1-339d-486f-8fe1-4670be8a14e7\" (UID: \"91f65ef1-339d-486f-8fe1-4670be8a14e7\") " Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.371247 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f65ef1-339d-486f-8fe1-4670be8a14e7-utilities\") pod \"91f65ef1-339d-486f-8fe1-4670be8a14e7\" (UID: \"91f65ef1-339d-486f-8fe1-4670be8a14e7\") " Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.374476 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91f65ef1-339d-486f-8fe1-4670be8a14e7-utilities" (OuterVolumeSpecName: "utilities") pod "91f65ef1-339d-486f-8fe1-4670be8a14e7" (UID: "91f65ef1-339d-486f-8fe1-4670be8a14e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.382396 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91f65ef1-339d-486f-8fe1-4670be8a14e7-kube-api-access-xczx6" (OuterVolumeSpecName: "kube-api-access-xczx6") pod "91f65ef1-339d-486f-8fe1-4670be8a14e7" (UID: "91f65ef1-339d-486f-8fe1-4670be8a14e7"). InnerVolumeSpecName "kube-api-access-xczx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.401803 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_5f4b4358-805a-4a55-a5a8-9200c0c5e5b5/cloudkitty-api/0.log" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.443545 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91f65ef1-339d-486f-8fe1-4670be8a14e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91f65ef1-339d-486f-8fe1-4670be8a14e7" (UID: "91f65ef1-339d-486f-8fe1-4670be8a14e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.474223 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f65ef1-339d-486f-8fe1-4670be8a14e7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.474274 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xczx6\" (UniqueName: \"kubernetes.io/projected/91f65ef1-339d-486f-8fe1-4670be8a14e7-kube-api-access-xczx6\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.474288 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f65ef1-339d-486f-8fe1-4670be8a14e7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.506917 4777 generic.go:334] "Generic (PLEG): container finished" podID="91f65ef1-339d-486f-8fe1-4670be8a14e7" containerID="c0845bf314c42044993e6aa74a3419bed10caffb54bd00ac54c380a45cd50582" exitCode=0 Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.506958 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jtbv" event={"ID":"91f65ef1-339d-486f-8fe1-4670be8a14e7","Type":"ContainerDied","Data":"c0845bf314c42044993e6aa74a3419bed10caffb54bd00ac54c380a45cd50582"} Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.506988 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jtbv" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.507009 4777 scope.go:117] "RemoveContainer" containerID="c0845bf314c42044993e6aa74a3419bed10caffb54bd00ac54c380a45cd50582" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.506996 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jtbv" event={"ID":"91f65ef1-339d-486f-8fe1-4670be8a14e7","Type":"ContainerDied","Data":"28bd6ce70d4653e493f82d1f39d6f8e6d473f694e0eaf8706f80f8208306ac76"} Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.525590 4777 scope.go:117] "RemoveContainer" containerID="09d712a531e3554fb4bb18b90ce8fb2fa4bf86ab9638118b719ffd0327b147f3" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.530955 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_5f4b4358-805a-4a55-a5a8-9200c0c5e5b5/cloudkitty-api-log/0.log" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.559732 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6jtbv"] Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.564180 4777 scope.go:117] "RemoveContainer" containerID="4b9d047d2dc2d3149912fbf4ec2c6f90ed23af4973c4f1109e1702b15e12b4bb" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.576592 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6jtbv"] Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.605649 4777 scope.go:117] "RemoveContainer" containerID="c0845bf314c42044993e6aa74a3419bed10caffb54bd00ac54c380a45cd50582" Nov 24 17:55:49 crc kubenswrapper[4777]: E1124 17:55:49.607383 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0845bf314c42044993e6aa74a3419bed10caffb54bd00ac54c380a45cd50582\": container with ID starting with c0845bf314c42044993e6aa74a3419bed10caffb54bd00ac54c380a45cd50582 not found: ID does not exist" containerID="c0845bf314c42044993e6aa74a3419bed10caffb54bd00ac54c380a45cd50582" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.607419 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0845bf314c42044993e6aa74a3419bed10caffb54bd00ac54c380a45cd50582"} err="failed to get container status \"c0845bf314c42044993e6aa74a3419bed10caffb54bd00ac54c380a45cd50582\": rpc error: code = NotFound desc = could not find container \"c0845bf314c42044993e6aa74a3419bed10caffb54bd00ac54c380a45cd50582\": container with ID starting with c0845bf314c42044993e6aa74a3419bed10caffb54bd00ac54c380a45cd50582 not found: ID does not exist" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.607440 4777 scope.go:117] "RemoveContainer" containerID="09d712a531e3554fb4bb18b90ce8fb2fa4bf86ab9638118b719ffd0327b147f3" Nov 24 17:55:49 crc kubenswrapper[4777]: E1124 17:55:49.608319 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09d712a531e3554fb4bb18b90ce8fb2fa4bf86ab9638118b719ffd0327b147f3\": container with ID starting with 09d712a531e3554fb4bb18b90ce8fb2fa4bf86ab9638118b719ffd0327b147f3 not found: ID does not exist" containerID="09d712a531e3554fb4bb18b90ce8fb2fa4bf86ab9638118b719ffd0327b147f3" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.608343 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09d712a531e3554fb4bb18b90ce8fb2fa4bf86ab9638118b719ffd0327b147f3"} err="failed to get container status \"09d712a531e3554fb4bb18b90ce8fb2fa4bf86ab9638118b719ffd0327b147f3\": rpc error: code = NotFound desc = could not find container \"09d712a531e3554fb4bb18b90ce8fb2fa4bf86ab9638118b719ffd0327b147f3\": container with ID starting with 09d712a531e3554fb4bb18b90ce8fb2fa4bf86ab9638118b719ffd0327b147f3 not found: ID does not exist" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.608355 4777 scope.go:117] "RemoveContainer" containerID="4b9d047d2dc2d3149912fbf4ec2c6f90ed23af4973c4f1109e1702b15e12b4bb" Nov 24 17:55:49 crc kubenswrapper[4777]: E1124 17:55:49.609825 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b9d047d2dc2d3149912fbf4ec2c6f90ed23af4973c4f1109e1702b15e12b4bb\": container with ID starting with 4b9d047d2dc2d3149912fbf4ec2c6f90ed23af4973c4f1109e1702b15e12b4bb not found: ID does not exist" containerID="4b9d047d2dc2d3149912fbf4ec2c6f90ed23af4973c4f1109e1702b15e12b4bb" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.609849 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b9d047d2dc2d3149912fbf4ec2c6f90ed23af4973c4f1109e1702b15e12b4bb"} err="failed to get container status \"4b9d047d2dc2d3149912fbf4ec2c6f90ed23af4973c4f1109e1702b15e12b4bb\": rpc error: code = NotFound desc = could not find container \"4b9d047d2dc2d3149912fbf4ec2c6f90ed23af4973c4f1109e1702b15e12b4bb\": container with ID starting with 4b9d047d2dc2d3149912fbf4ec2c6f90ed23af4973c4f1109e1702b15e12b4bb not found: ID does not exist" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.725174 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-distributor-56cd74f89f-jg9l8_9c73a207-d21c-4157-b27a-41656a5f6af1/loki-distributor/0.log" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.729002 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-compactor-0_e99dec76-e21f-4aab-bdc9-ae098391ee6e/loki-compactor/0.log" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.918810 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-76cc998948-qskdh_f1facf54-ce8f-441b-bdef-d864d3bc5dd7/gateway/0.log" Nov 24 17:55:49 crc kubenswrapper[4777]: I1124 17:55:49.993706 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-76cc998948-znzk8_1aec5e66-69c2-404c-865a-0995f09fd435/gateway/0.log" Nov 24 17:55:50 crc kubenswrapper[4777]: I1124 17:55:50.320646 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-index-gateway-0_8ddeb3e7-8676-40b5-918b-6d8ac12a2b35/loki-index-gateway/0.log" Nov 24 17:55:50 crc kubenswrapper[4777]: I1124 17:55:50.388073 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-ingester-0_f68bfbdc-e400-488d-9f32-ea90fd1160e3/loki-ingester/0.log" Nov 24 17:55:50 crc kubenswrapper[4777]: I1124 17:55:50.737683 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-query-frontend-779849886d-x7zhc_0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83/loki-query-frontend/0.log" Nov 24 17:55:50 crc kubenswrapper[4777]: I1124 17:55:50.989927 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-querier-548665d79b-tqjct_c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb/loki-querier/0.log" Nov 24 17:55:51 crc kubenswrapper[4777]: I1124 17:55:51.021450 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz_45884cb1-7d6c-47a3-b64f-939c1119d70b/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:55:51 crc kubenswrapper[4777]: I1124 17:55:51.258411 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91f65ef1-339d-486f-8fe1-4670be8a14e7" path="/var/lib/kubelet/pods/91f65ef1-339d-486f-8fe1-4670be8a14e7/volumes" Nov 24 17:55:51 crc kubenswrapper[4777]: I1124 17:55:51.526138 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-qp66f_b21ec255-1398-4613-b461-29fd5ce00ab1/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:55:51 crc kubenswrapper[4777]: I1124 17:55:51.582560 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c4b758ff5-frg8p_33b50ce8-58e3-457c-81ce-dde13962d338/init/0.log" Nov 24 17:55:52 crc kubenswrapper[4777]: I1124 17:55:52.033684 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c4b758ff5-frg8p_33b50ce8-58e3-457c-81ce-dde13962d338/init/0.log" Nov 24 17:55:52 crc kubenswrapper[4777]: I1124 17:55:52.048571 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c4b758ff5-frg8p_33b50ce8-58e3-457c-81ce-dde13962d338/dnsmasq-dns/0.log" Nov 24 17:55:52 crc kubenswrapper[4777]: I1124 17:55:52.119092 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j_266eceac-bfd7-4597-b218-62cea462ce5c/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:55:52 crc kubenswrapper[4777]: I1124 17:55:52.296609 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_44214c39-ae9a-4e14-a7ec-fc1e95051d7f/glance-httpd/0.log" Nov 24 17:55:52 crc kubenswrapper[4777]: I1124 17:55:52.301260 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_44214c39-ae9a-4e14-a7ec-fc1e95051d7f/glance-log/0.log" Nov 24 17:55:52 crc kubenswrapper[4777]: I1124 17:55:52.492229 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b698d8b2-ca7b-43e7-8da9-aef7d3ba1538/glance-httpd/0.log" Nov 24 17:55:52 crc kubenswrapper[4777]: I1124 17:55:52.551199 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b698d8b2-ca7b-43e7-8da9-aef7d3ba1538/glance-log/0.log" Nov 24 17:55:52 crc kubenswrapper[4777]: I1124 17:55:52.829319 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw_0a98c33f-1b9f-4419-8dd3-ea3ba5695980/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:55:52 crc kubenswrapper[4777]: I1124 17:55:52.930380 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-rvlp6_346142b6-2a63-4610-b1d0-bfebac8c7c61/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:55:53 crc kubenswrapper[4777]: I1124 17:55:53.195133 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_1e1ab5a6-3bb5-4722-9220-1a033a656b93/kube-state-metrics/0.log" Nov 24 17:55:53 crc kubenswrapper[4777]: I1124 17:55:53.465258 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-md792_fd79eb9e-a6a2-4932-8a81-27a115ccaef6/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:55:53 crc kubenswrapper[4777]: I1124 17:55:53.737352 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5877679d68-q5nlj_8021c40d-848c-4365-84f2-9878b03d4656/keystone-api/0.log" Nov 24 17:55:54 crc kubenswrapper[4777]: I1124 17:55:54.055851 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5984748495-2rxn6_42e06676-5c5b-4684-b525-669d27d31818/neutron-api/0.log" Nov 24 17:55:54 crc kubenswrapper[4777]: I1124 17:55:54.072287 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5984748495-2rxn6_42e06676-5c5b-4684-b525-669d27d31818/neutron-httpd/0.log" Nov 24 17:55:54 crc kubenswrapper[4777]: I1124 17:55:54.335558 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2_97805347-c06c-409d-a299-cd4cfdf7e98e/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:55:54 crc kubenswrapper[4777]: I1124 17:55:54.903630 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_f474cf0f-37fd-495d-ac31-493641c222a7/nova-api-log/0.log" Nov 24 17:55:54 crc kubenswrapper[4777]: I1124 17:55:54.960089 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_f474cf0f-37fd-495d-ac31-493641c222a7/nova-api-api/0.log" Nov 24 17:55:55 crc kubenswrapper[4777]: I1124 17:55:55.261836 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_aee763a5-b286-4fdd-9573-d06a0e915f9e/nova-cell0-conductor-conductor/0.log" Nov 24 17:55:55 crc kubenswrapper[4777]: I1124 17:55:55.622767 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b929fcfa-b979-4f40-a469-fdd1e24a9ef2/nova-cell1-conductor-conductor/0.log" Nov 24 17:55:55 crc kubenswrapper[4777]: I1124 17:55:55.772983 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_ad8c715b-ea28-4366-882c-4371692973f0/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 17:55:55 crc kubenswrapper[4777]: I1124 17:55:55.914092 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-w44w5_9651259d-da1a-4580-855a-94a9310fc952/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:55:56 crc kubenswrapper[4777]: I1124 17:55:56.248735 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4da8a01e-98cb-4d18-b86e-c281010dcf8e/nova-metadata-log/0.log" Nov 24 17:55:56 crc kubenswrapper[4777]: I1124 17:55:56.258981 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-proc-0_1c366c57-2d20-4e76-b90f-c90fc01347fc/cloudkitty-proc/0.log" Nov 24 17:55:56 crc kubenswrapper[4777]: I1124 17:55:56.718736 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_2a63c908-3ee5-449c-ab1c-6ca209338bee/nova-scheduler-scheduler/0.log" Nov 24 17:55:56 crc kubenswrapper[4777]: I1124 17:55:56.840620 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4eba3f22-b7ad-464d-a0c9-437c449d124b/mysql-bootstrap/0.log" Nov 24 17:55:57 crc kubenswrapper[4777]: I1124 17:55:57.082191 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4eba3f22-b7ad-464d-a0c9-437c449d124b/mysql-bootstrap/0.log" Nov 24 17:55:57 crc kubenswrapper[4777]: I1124 17:55:57.146914 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4eba3f22-b7ad-464d-a0c9-437c449d124b/galera/0.log" Nov 24 17:55:57 crc kubenswrapper[4777]: I1124 17:55:57.197407 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4da8a01e-98cb-4d18-b86e-c281010dcf8e/nova-metadata-metadata/0.log" Nov 24 17:55:57 crc kubenswrapper[4777]: I1124 17:55:57.370777 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1d71a086-37e0-4409-9db9-66e6f1374406/mysql-bootstrap/0.log" Nov 24 17:55:57 crc kubenswrapper[4777]: I1124 17:55:57.591704 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1d71a086-37e0-4409-9db9-66e6f1374406/galera/0.log" Nov 24 17:55:57 crc kubenswrapper[4777]: I1124 17:55:57.629203 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1d71a086-37e0-4409-9db9-66e6f1374406/mysql-bootstrap/0.log" Nov 24 17:55:57 crc kubenswrapper[4777]: I1124 17:55:57.677394 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2d497520-8510-44d4-b57c-9b8e35ebaa32/openstackclient/0.log" Nov 24 17:55:57 crc kubenswrapper[4777]: I1124 17:55:57.859202 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-lgxzj_04bed066-31b6-49d0-90ee-68e38f7944c1/openstack-network-exporter/0.log" Nov 24 17:55:58 crc kubenswrapper[4777]: I1124 17:55:58.043300 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dkbds_3641f728-c179-4f10-93c2-98f74763e77f/ovsdb-server-init/0.log" Nov 24 17:55:58 crc kubenswrapper[4777]: I1124 17:55:58.310122 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dkbds_3641f728-c179-4f10-93c2-98f74763e77f/ovsdb-server-init/0.log" Nov 24 17:55:58 crc kubenswrapper[4777]: I1124 17:55:58.364283 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dkbds_3641f728-c179-4f10-93c2-98f74763e77f/ovsdb-server/0.log" Nov 24 17:55:58 crc kubenswrapper[4777]: I1124 17:55:58.418252 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dkbds_3641f728-c179-4f10-93c2-98f74763e77f/ovs-vswitchd/0.log" Nov 24 17:55:58 crc kubenswrapper[4777]: I1124 17:55:58.541375 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fm4mt" podUID="7e418e82-f285-4adf-a6ac-e425aa66c9c8" containerName="registry-server" probeResult="failure" output=< Nov 24 17:55:58 crc kubenswrapper[4777]: timeout: failed to connect service ":50051" within 1s Nov 24 17:55:58 crc kubenswrapper[4777]: > Nov 24 17:55:58 crc kubenswrapper[4777]: I1124 17:55:58.627880 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-q4l2k_73928e73-667c-4b69-aed9-72f32012fbdc/ovn-controller/0.log" Nov 24 17:55:58 crc kubenswrapper[4777]: I1124 17:55:58.683164 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-xkd2c_9c522429-96d7-44c9-812c-f86726228305/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:55:58 crc kubenswrapper[4777]: I1124 17:55:58.942554 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_001aba35-921b-43b5-87e4-7e92d5595a7b/openstack-network-exporter/0.log" Nov 24 17:55:58 crc kubenswrapper[4777]: I1124 17:55:58.943849 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_001aba35-921b-43b5-87e4-7e92d5595a7b/ovn-northd/0.log" Nov 24 17:55:59 crc kubenswrapper[4777]: I1124 17:55:59.050982 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4ccca46a-1cac-4693-8490-2612b138c9d5/openstack-network-exporter/0.log" Nov 24 17:55:59 crc kubenswrapper[4777]: I1124 17:55:59.269529 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4ccca46a-1cac-4693-8490-2612b138c9d5/ovsdbserver-nb/0.log" Nov 24 17:55:59 crc kubenswrapper[4777]: I1124 17:55:59.277438 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_737d5505-10bc-4e56-b846-01a3af071b38/ovsdbserver-sb/0.log" Nov 24 17:55:59 crc kubenswrapper[4777]: I1124 17:55:59.308123 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_737d5505-10bc-4e56-b846-01a3af071b38/openstack-network-exporter/0.log" Nov 24 17:55:59 crc kubenswrapper[4777]: I1124 17:55:59.621293 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-55f98d7b-5njfr_b25fa9b0-aec5-4a33-aefe-2ee5685d0e88/placement-api/0.log" Nov 24 17:55:59 crc kubenswrapper[4777]: I1124 17:55:59.728840 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-55f98d7b-5njfr_b25fa9b0-aec5-4a33-aefe-2ee5685d0e88/placement-log/0.log" Nov 24 17:55:59 crc kubenswrapper[4777]: I1124 17:55:59.967136 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_df409568-e0d5-49db-8da4-efbe45ec3a2c/init-config-reloader/0.log" Nov 24 17:56:00 crc kubenswrapper[4777]: I1124 17:56:00.172930 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_df409568-e0d5-49db-8da4-efbe45ec3a2c/init-config-reloader/0.log" Nov 24 17:56:00 crc kubenswrapper[4777]: I1124 17:56:00.212937 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_df409568-e0d5-49db-8da4-efbe45ec3a2c/thanos-sidecar/0.log" Nov 24 17:56:00 crc kubenswrapper[4777]: I1124 17:56:00.239093 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_df409568-e0d5-49db-8da4-efbe45ec3a2c/config-reloader/0.log" Nov 24 17:56:00 crc kubenswrapper[4777]: I1124 17:56:00.277525 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_df409568-e0d5-49db-8da4-efbe45ec3a2c/prometheus/0.log" Nov 24 17:56:00 crc kubenswrapper[4777]: I1124 17:56:00.399179 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_42d2ab97-ca44-4234-8405-22f8797e0c0b/setup-container/0.log" Nov 24 17:56:00 crc kubenswrapper[4777]: I1124 17:56:00.774086 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_42d2ab97-ca44-4234-8405-22f8797e0c0b/rabbitmq/0.log" Nov 24 17:56:00 crc kubenswrapper[4777]: I1124 17:56:00.786910 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_42d2ab97-ca44-4234-8405-22f8797e0c0b/setup-container/0.log" Nov 24 17:56:00 crc kubenswrapper[4777]: I1124 17:56:00.847568 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_78712b89-27c4-4481-b152-320fb7fee9fc/setup-container/0.log" Nov 24 17:56:01 crc kubenswrapper[4777]: I1124 17:56:01.042889 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_78712b89-27c4-4481-b152-320fb7fee9fc/setup-container/0.log" Nov 24 17:56:01 crc kubenswrapper[4777]: I1124 17:56:01.059020 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_78712b89-27c4-4481-b152-320fb7fee9fc/rabbitmq/0.log" Nov 24 17:56:01 crc kubenswrapper[4777]: I1124 17:56:01.171103 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6_902e59e0-e5a3-4f96-953e-ec961dbe53d9/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:56:01 crc kubenswrapper[4777]: I1124 17:56:01.310160 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-qfmq5_f27aca44-4a34-469c-97fd-075eed1300dd/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:56:01 crc kubenswrapper[4777]: I1124 17:56:01.441535 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq_10de4ea3-2dc3-4e57-ac55-d6be12743347/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:56:01 crc kubenswrapper[4777]: I1124 17:56:01.632499 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-kq565_a57c6a80-0881-430c-a6bb-beea385e0417/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:56:01 crc kubenswrapper[4777]: I1124 17:56:01.735512 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-q4nrb_3be75c8d-7695-403e-b801-d2650e0b331c/ssh-known-hosts-edpm-deployment/0.log" Nov 24 17:56:01 crc kubenswrapper[4777]: I1124 17:56:01.973318 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-644f75df95-wzmtb_22f2176a-dcae-476f-92b5-eb3d19d9bb4f/proxy-server/0.log" Nov 24 17:56:02 crc kubenswrapper[4777]: I1124 17:56:02.043308 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-644f75df95-wzmtb_22f2176a-dcae-476f-92b5-eb3d19d9bb4f/proxy-httpd/0.log" Nov 24 17:56:02 crc kubenswrapper[4777]: I1124 17:56:02.110100 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-xgm6s_3ec24d0c-556b-4a24-8f42-38fed08efe6b/swift-ring-rebalance/0.log" Nov 24 17:56:02 crc kubenswrapper[4777]: I1124 17:56:02.300458 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/account-reaper/0.log" Nov 24 17:56:02 crc kubenswrapper[4777]: I1124 17:56:02.335838 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/account-auditor/0.log" Nov 24 17:56:02 crc kubenswrapper[4777]: I1124 17:56:02.531380 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/account-replicator/0.log" Nov 24 17:56:02 crc kubenswrapper[4777]: I1124 17:56:02.552760 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/container-auditor/0.log" Nov 24 17:56:02 crc kubenswrapper[4777]: I1124 17:56:02.555354 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/account-server/0.log" Nov 24 17:56:02 crc kubenswrapper[4777]: I1124 17:56:02.690945 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/container-replicator/0.log" Nov 24 17:56:02 crc kubenswrapper[4777]: I1124 17:56:02.729019 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/container-server/0.log" Nov 24 17:56:02 crc kubenswrapper[4777]: I1124 17:56:02.832553 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/container-updater/0.log" Nov 24 17:56:02 crc kubenswrapper[4777]: I1124 17:56:02.927857 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/object-auditor/0.log" Nov 24 17:56:02 crc kubenswrapper[4777]: I1124 17:56:02.975437 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/object-expirer/0.log" Nov 24 17:56:03 crc kubenswrapper[4777]: I1124 17:56:03.009745 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/object-replicator/0.log" Nov 24 17:56:03 crc kubenswrapper[4777]: I1124 17:56:03.162823 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/object-server/0.log" Nov 24 17:56:03 crc kubenswrapper[4777]: I1124 17:56:03.270788 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/object-updater/0.log" Nov 24 17:56:03 crc kubenswrapper[4777]: I1124 17:56:03.488210 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/rsync/0.log" Nov 24 17:56:03 crc kubenswrapper[4777]: I1124 17:56:03.564208 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/swift-recon-cron/0.log" Nov 24 17:56:03 crc kubenswrapper[4777]: I1124 17:56:03.638529 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl_bcd7d063-a7cc-4f12-8ba2-9e7578152a3e/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:56:03 crc kubenswrapper[4777]: I1124 17:56:03.828259 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_08d58d9b-d4f1-458d-bfe3-fa8bc585f352/tempest-tests-tempest-tests-runner/0.log" Nov 24 17:56:03 crc kubenswrapper[4777]: I1124 17:56:03.887732 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_2d149ad6-8d86-445c-baed-c43f1eaec2cb/test-operator-logs-container/0.log" Nov 24 17:56:04 crc kubenswrapper[4777]: I1124 17:56:04.134592 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j_541aa87c-3013-4641-950b-a23f825bcb96/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 17:56:07 crc kubenswrapper[4777]: I1124 17:56:07.829592 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_0254e0e3-f9b9-4017-954a-14d3d21e37c5/memcached/0.log" Nov 24 17:56:08 crc kubenswrapper[4777]: I1124 17:56:08.561017 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fm4mt" podUID="7e418e82-f285-4adf-a6ac-e425aa66c9c8" containerName="registry-server" probeResult="failure" output=< Nov 24 17:56:08 crc kubenswrapper[4777]: timeout: failed to connect service ":50051" within 1s Nov 24 17:56:08 crc kubenswrapper[4777]: > Nov 24 17:56:17 crc kubenswrapper[4777]: I1124 17:56:17.555601 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:56:17 crc kubenswrapper[4777]: I1124 17:56:17.622196 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:56:17 crc kubenswrapper[4777]: I1124 17:56:17.793938 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fm4mt"] Nov 24 17:56:18 crc kubenswrapper[4777]: I1124 17:56:18.821642 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fm4mt" podUID="7e418e82-f285-4adf-a6ac-e425aa66c9c8" containerName="registry-server" containerID="cri-o://7a9fec5b3da6967d367fbf946d5ccc54ab9fc21561ae7675a4fbc5ced5d7d426" gracePeriod=2 Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.622890 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.720521 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4646r\" (UniqueName: \"kubernetes.io/projected/7e418e82-f285-4adf-a6ac-e425aa66c9c8-kube-api-access-4646r\") pod \"7e418e82-f285-4adf-a6ac-e425aa66c9c8\" (UID: \"7e418e82-f285-4adf-a6ac-e425aa66c9c8\") " Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.720709 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e418e82-f285-4adf-a6ac-e425aa66c9c8-catalog-content\") pod \"7e418e82-f285-4adf-a6ac-e425aa66c9c8\" (UID: \"7e418e82-f285-4adf-a6ac-e425aa66c9c8\") " Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.720878 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e418e82-f285-4adf-a6ac-e425aa66c9c8-utilities\") pod \"7e418e82-f285-4adf-a6ac-e425aa66c9c8\" (UID: \"7e418e82-f285-4adf-a6ac-e425aa66c9c8\") " Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.722158 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e418e82-f285-4adf-a6ac-e425aa66c9c8-utilities" (OuterVolumeSpecName: "utilities") pod "7e418e82-f285-4adf-a6ac-e425aa66c9c8" (UID: "7e418e82-f285-4adf-a6ac-e425aa66c9c8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.740160 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e418e82-f285-4adf-a6ac-e425aa66c9c8-kube-api-access-4646r" (OuterVolumeSpecName: "kube-api-access-4646r") pod "7e418e82-f285-4adf-a6ac-e425aa66c9c8" (UID: "7e418e82-f285-4adf-a6ac-e425aa66c9c8"). InnerVolumeSpecName "kube-api-access-4646r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.812747 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e418e82-f285-4adf-a6ac-e425aa66c9c8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e418e82-f285-4adf-a6ac-e425aa66c9c8" (UID: "7e418e82-f285-4adf-a6ac-e425aa66c9c8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.823577 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e418e82-f285-4adf-a6ac-e425aa66c9c8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.823646 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e418e82-f285-4adf-a6ac-e425aa66c9c8-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.823659 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4646r\" (UniqueName: \"kubernetes.io/projected/7e418e82-f285-4adf-a6ac-e425aa66c9c8-kube-api-access-4646r\") on node \"crc\" DevicePath \"\"" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.835672 4777 generic.go:334] "Generic (PLEG): container finished" podID="7e418e82-f285-4adf-a6ac-e425aa66c9c8" containerID="7a9fec5b3da6967d367fbf946d5ccc54ab9fc21561ae7675a4fbc5ced5d7d426" exitCode=0 Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.835720 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fm4mt" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.835735 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fm4mt" event={"ID":"7e418e82-f285-4adf-a6ac-e425aa66c9c8","Type":"ContainerDied","Data":"7a9fec5b3da6967d367fbf946d5ccc54ab9fc21561ae7675a4fbc5ced5d7d426"} Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.835777 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fm4mt" event={"ID":"7e418e82-f285-4adf-a6ac-e425aa66c9c8","Type":"ContainerDied","Data":"6a0e1b5eb4ccda19a255d236d5a900238fca67d56560f5f15f91398b0b840388"} Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.835805 4777 scope.go:117] "RemoveContainer" containerID="7a9fec5b3da6967d367fbf946d5ccc54ab9fc21561ae7675a4fbc5ced5d7d426" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.872452 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fm4mt"] Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.872739 4777 scope.go:117] "RemoveContainer" containerID="c3232a1911fb7751da9aeb7ccdd79d724709b42ac0d9a6b99b3a071c69a3b4a2" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.883423 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fm4mt"] Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.906665 4777 scope.go:117] "RemoveContainer" containerID="c180bbeef8ff04fde81dd992ed00d37aab8f5778c2a9750bd97c9323d0c76717" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.979197 4777 scope.go:117] "RemoveContainer" containerID="7a9fec5b3da6967d367fbf946d5ccc54ab9fc21561ae7675a4fbc5ced5d7d426" Nov 24 17:56:19 crc kubenswrapper[4777]: E1124 17:56:19.984070 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a9fec5b3da6967d367fbf946d5ccc54ab9fc21561ae7675a4fbc5ced5d7d426\": container with ID starting with 7a9fec5b3da6967d367fbf946d5ccc54ab9fc21561ae7675a4fbc5ced5d7d426 not found: ID does not exist" containerID="7a9fec5b3da6967d367fbf946d5ccc54ab9fc21561ae7675a4fbc5ced5d7d426" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.984114 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a9fec5b3da6967d367fbf946d5ccc54ab9fc21561ae7675a4fbc5ced5d7d426"} err="failed to get container status \"7a9fec5b3da6967d367fbf946d5ccc54ab9fc21561ae7675a4fbc5ced5d7d426\": rpc error: code = NotFound desc = could not find container \"7a9fec5b3da6967d367fbf946d5ccc54ab9fc21561ae7675a4fbc5ced5d7d426\": container with ID starting with 7a9fec5b3da6967d367fbf946d5ccc54ab9fc21561ae7675a4fbc5ced5d7d426 not found: ID does not exist" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.984140 4777 scope.go:117] "RemoveContainer" containerID="c3232a1911fb7751da9aeb7ccdd79d724709b42ac0d9a6b99b3a071c69a3b4a2" Nov 24 17:56:19 crc kubenswrapper[4777]: E1124 17:56:19.985188 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3232a1911fb7751da9aeb7ccdd79d724709b42ac0d9a6b99b3a071c69a3b4a2\": container with ID starting with c3232a1911fb7751da9aeb7ccdd79d724709b42ac0d9a6b99b3a071c69a3b4a2 not found: ID does not exist" containerID="c3232a1911fb7751da9aeb7ccdd79d724709b42ac0d9a6b99b3a071c69a3b4a2" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.985215 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3232a1911fb7751da9aeb7ccdd79d724709b42ac0d9a6b99b3a071c69a3b4a2"} err="failed to get container status \"c3232a1911fb7751da9aeb7ccdd79d724709b42ac0d9a6b99b3a071c69a3b4a2\": rpc error: code = NotFound desc = could not find container \"c3232a1911fb7751da9aeb7ccdd79d724709b42ac0d9a6b99b3a071c69a3b4a2\": container with ID starting with c3232a1911fb7751da9aeb7ccdd79d724709b42ac0d9a6b99b3a071c69a3b4a2 not found: ID does not exist" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.985232 4777 scope.go:117] "RemoveContainer" containerID="c180bbeef8ff04fde81dd992ed00d37aab8f5778c2a9750bd97c9323d0c76717" Nov 24 17:56:19 crc kubenswrapper[4777]: E1124 17:56:19.990897 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c180bbeef8ff04fde81dd992ed00d37aab8f5778c2a9750bd97c9323d0c76717\": container with ID starting with c180bbeef8ff04fde81dd992ed00d37aab8f5778c2a9750bd97c9323d0c76717 not found: ID does not exist" containerID="c180bbeef8ff04fde81dd992ed00d37aab8f5778c2a9750bd97c9323d0c76717" Nov 24 17:56:19 crc kubenswrapper[4777]: I1124 17:56:19.990944 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c180bbeef8ff04fde81dd992ed00d37aab8f5778c2a9750bd97c9323d0c76717"} err="failed to get container status \"c180bbeef8ff04fde81dd992ed00d37aab8f5778c2a9750bd97c9323d0c76717\": rpc error: code = NotFound desc = could not find container \"c180bbeef8ff04fde81dd992ed00d37aab8f5778c2a9750bd97c9323d0c76717\": container with ID starting with c180bbeef8ff04fde81dd992ed00d37aab8f5778c2a9750bd97c9323d0c76717 not found: ID does not exist" Nov 24 17:56:21 crc kubenswrapper[4777]: I1124 17:56:21.256922 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e418e82-f285-4adf-a6ac-e425aa66c9c8" path="/var/lib/kubelet/pods/7e418e82-f285-4adf-a6ac-e425aa66c9c8/volumes" Nov 24 17:56:30 crc kubenswrapper[4777]: I1124 17:56:30.629993 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl_bd350075-1ebb-4431-9b02-91e2d6c5e7d3/util/0.log" Nov 24 17:56:30 crc kubenswrapper[4777]: I1124 17:56:30.860961 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl_bd350075-1ebb-4431-9b02-91e2d6c5e7d3/pull/0.log" Nov 24 17:56:30 crc kubenswrapper[4777]: I1124 17:56:30.924524 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl_bd350075-1ebb-4431-9b02-91e2d6c5e7d3/pull/0.log" Nov 24 17:56:30 crc kubenswrapper[4777]: I1124 17:56:30.936576 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl_bd350075-1ebb-4431-9b02-91e2d6c5e7d3/util/0.log" Nov 24 17:56:31 crc kubenswrapper[4777]: I1124 17:56:31.024148 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl_bd350075-1ebb-4431-9b02-91e2d6c5e7d3/util/0.log" Nov 24 17:56:31 crc kubenswrapper[4777]: I1124 17:56:31.052320 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl_bd350075-1ebb-4431-9b02-91e2d6c5e7d3/pull/0.log" Nov 24 17:56:31 crc kubenswrapper[4777]: I1124 17:56:31.131516 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl_bd350075-1ebb-4431-9b02-91e2d6c5e7d3/extract/0.log" Nov 24 17:56:31 crc kubenswrapper[4777]: I1124 17:56:31.245448 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-82ppc_a7457835-23e4-460e-a155-07c1d4e7c30e/kube-rbac-proxy/0.log" Nov 24 17:56:31 crc kubenswrapper[4777]: I1124 17:56:31.367810 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-82ppc_a7457835-23e4-460e-a155-07c1d4e7c30e/manager/0.log" Nov 24 17:56:31 crc kubenswrapper[4777]: I1124 17:56:31.398840 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-2chns_1d2ecc53-902f-41f3-80f4-93f9e48c8532/kube-rbac-proxy/0.log" Nov 24 17:56:31 crc kubenswrapper[4777]: I1124 17:56:31.554591 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-2chns_1d2ecc53-902f-41f3-80f4-93f9e48c8532/manager/0.log" Nov 24 17:56:31 crc kubenswrapper[4777]: I1124 17:56:31.615940 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-klpxr_99bfc986-b80d-4193-bf32-4488c11f0066/kube-rbac-proxy/0.log" Nov 24 17:56:31 crc kubenswrapper[4777]: I1124 17:56:31.640860 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-klpxr_99bfc986-b80d-4193-bf32-4488c11f0066/manager/0.log" Nov 24 17:56:31 crc kubenswrapper[4777]: I1124 17:56:31.819269 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-9mzvw_fe817743-7f1e-46a4-9757-5b678ddc097a/kube-rbac-proxy/0.log" Nov 24 17:56:31 crc kubenswrapper[4777]: I1124 17:56:31.951559 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-9mzvw_fe817743-7f1e-46a4-9757-5b678ddc097a/manager/0.log" Nov 24 17:56:32 crc kubenswrapper[4777]: I1124 17:56:32.039239 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-zmms9_1d3fc8a3-f3cf-4f37-846d-b244a52415f8/kube-rbac-proxy/0.log" Nov 24 17:56:32 crc kubenswrapper[4777]: I1124 17:56:32.072467 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-zmms9_1d3fc8a3-f3cf-4f37-846d-b244a52415f8/manager/0.log" Nov 24 17:56:32 crc kubenswrapper[4777]: I1124 17:56:32.128007 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-6xxk9_8401afce-ddb4-4195-b6ae-7e5c91128525/kube-rbac-proxy/0.log" Nov 24 17:56:32 crc kubenswrapper[4777]: I1124 17:56:32.261729 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-6xxk9_8401afce-ddb4-4195-b6ae-7e5c91128525/manager/0.log" Nov 24 17:56:32 crc kubenswrapper[4777]: I1124 17:56:32.349334 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-r5pxc_b755cc8b-dd00-4b28-8fb4-908e0e3db9a5/kube-rbac-proxy/0.log" Nov 24 17:56:32 crc kubenswrapper[4777]: I1124 17:56:32.471692 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-r5pxc_b755cc8b-dd00-4b28-8fb4-908e0e3db9a5/manager/0.log" Nov 24 17:56:32 crc kubenswrapper[4777]: I1124 17:56:32.504942 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-wmvls_f7f44c61-027a-41c4-9665-8394d579ba33/kube-rbac-proxy/0.log" Nov 24 17:56:32 crc kubenswrapper[4777]: I1124 17:56:32.634877 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-wmvls_f7f44c61-027a-41c4-9665-8394d579ba33/manager/0.log" Nov 24 17:56:32 crc kubenswrapper[4777]: I1124 17:56:32.688257 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-v5vgx_f1425713-2fd6-4108-abe3-9f46063e98f5/kube-rbac-proxy/0.log" Nov 24 17:56:32 crc kubenswrapper[4777]: I1124 17:56:32.803810 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-v5vgx_f1425713-2fd6-4108-abe3-9f46063e98f5/manager/0.log" Nov 24 17:56:32 crc kubenswrapper[4777]: I1124 17:56:32.932739 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-mf8r8_38f287c7-6c43-4dc9-b547-0d494704874a/kube-rbac-proxy/0.log" Nov 24 17:56:32 crc kubenswrapper[4777]: I1124 17:56:32.997882 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-mf8r8_38f287c7-6c43-4dc9-b547-0d494704874a/manager/0.log" Nov 24 17:56:33 crc kubenswrapper[4777]: I1124 17:56:33.113384 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-g9kpn_73a88858-0760-4b40-b57f-71e4f9977129/kube-rbac-proxy/0.log" Nov 24 17:56:33 crc kubenswrapper[4777]: I1124 17:56:33.180954 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-g9kpn_73a88858-0760-4b40-b57f-71e4f9977129/manager/0.log" Nov 24 17:56:33 crc kubenswrapper[4777]: I1124 17:56:33.245193 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-dc68g_fd1db7a5-22e3-4f16-a026-0b8895ef292d/kube-rbac-proxy/0.log" Nov 24 17:56:33 crc kubenswrapper[4777]: I1124 17:56:33.399604 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-dc68g_fd1db7a5-22e3-4f16-a026-0b8895ef292d/manager/0.log" Nov 24 17:56:33 crc kubenswrapper[4777]: I1124 17:56:33.461779 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-58vj4_a9a5349a-868d-4c3e-a3ab-f57b55643759/kube-rbac-proxy/0.log" Nov 24 17:56:33 crc kubenswrapper[4777]: I1124 17:56:33.559073 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-58vj4_a9a5349a-868d-4c3e-a3ab-f57b55643759/manager/0.log" Nov 24 17:56:33 crc kubenswrapper[4777]: I1124 17:56:33.619961 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-ss948_7ea09c1e-d85f-435e-a029-8b4f0df06839/kube-rbac-proxy/0.log" Nov 24 17:56:33 crc kubenswrapper[4777]: I1124 17:56:33.710415 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-ss948_7ea09c1e-d85f-435e-a029-8b4f0df06839/manager/0.log" Nov 24 17:56:33 crc kubenswrapper[4777]: I1124 17:56:33.857955 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d_4a491c1c-85db-4efd-b944-40b1651a3c18/kube-rbac-proxy/0.log" Nov 24 17:56:33 crc kubenswrapper[4777]: I1124 17:56:33.889517 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d_4a491c1c-85db-4efd-b944-40b1651a3c18/manager/0.log" Nov 24 17:56:34 crc kubenswrapper[4777]: I1124 17:56:34.073751 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7856bf9fdc-k8wq8_783eef36-329f-4ceb-b516-020fe4cf5151/kube-rbac-proxy/0.log" Nov 24 17:56:34 crc kubenswrapper[4777]: I1124 17:56:34.257837 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-68bb788c6f-ft4rz_547759be-2d04-4553-aedc-e3f1b7ef9779/kube-rbac-proxy/0.log" Nov 24 17:56:34 crc kubenswrapper[4777]: I1124 17:56:34.469890 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zvxss_127dc4f2-329f-453e-8f21-1877225cec81/registry-server/0.log" Nov 24 17:56:34 crc kubenswrapper[4777]: I1124 17:56:34.484828 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-68bb788c6f-ft4rz_547759be-2d04-4553-aedc-e3f1b7ef9779/operator/0.log" Nov 24 17:56:34 crc kubenswrapper[4777]: I1124 17:56:34.644060 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-vxgn2_cfd203de-6a3e-4fdf-af55-3f5eaa62c5c7/kube-rbac-proxy/0.log" Nov 24 17:56:34 crc kubenswrapper[4777]: I1124 17:56:34.709602 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-hmsp7_ef1391a6-60c0-47c9-aa4f-7fa39f90d55b/kube-rbac-proxy/0.log" Nov 24 17:56:34 crc kubenswrapper[4777]: I1124 17:56:34.767923 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-vxgn2_cfd203de-6a3e-4fdf-af55-3f5eaa62c5c7/manager/0.log" Nov 24 17:56:34 crc kubenswrapper[4777]: I1124 17:56:34.948250 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-hmsp7_ef1391a6-60c0-47c9-aa4f-7fa39f90d55b/manager/0.log" Nov 24 17:56:35 crc kubenswrapper[4777]: I1124 17:56:35.026790 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm_386e5031-5136-4c3f-9ba2-6051496e0ed9/operator/0.log" Nov 24 17:56:35 crc kubenswrapper[4777]: I1124 17:56:35.186384 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7856bf9fdc-k8wq8_783eef36-329f-4ceb-b516-020fe4cf5151/manager/0.log" Nov 24 17:56:35 crc kubenswrapper[4777]: I1124 17:56:35.282612 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-q2p2n_c847a619-1224-4a3c-bc57-bb75924d8553/manager/0.log" Nov 24 17:56:35 crc kubenswrapper[4777]: I1124 17:56:35.339351 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-q2p2n_c847a619-1224-4a3c-bc57-bb75924d8553/kube-rbac-proxy/0.log" Nov 24 17:56:35 crc kubenswrapper[4777]: I1124 17:56:35.430614 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-64dc7b4899-mmn8p_6f802034-e827-4870-9e32-d1d0cc263587/kube-rbac-proxy/0.log" Nov 24 17:56:35 crc kubenswrapper[4777]: I1124 17:56:35.557898 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-jbqdt_2f5425e7-4f78-4b35-bb3f-b1c0065cfc42/kube-rbac-proxy/0.log" Nov 24 17:56:35 crc kubenswrapper[4777]: I1124 17:56:35.589113 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-jbqdt_2f5425e7-4f78-4b35-bb3f-b1c0065cfc42/manager/0.log" Nov 24 17:56:35 crc kubenswrapper[4777]: I1124 17:56:35.793587 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-64dc7b4899-mmn8p_6f802034-e827-4870-9e32-d1d0cc263587/manager/0.log" Nov 24 17:56:35 crc kubenswrapper[4777]: I1124 17:56:35.794997 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-g5jdq_44e68f24-0747-476d-b3e5-11e2952c3792/kube-rbac-proxy/0.log" Nov 24 17:56:35 crc kubenswrapper[4777]: I1124 17:56:35.830136 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-g5jdq_44e68f24-0747-476d-b3e5-11e2952c3792/manager/0.log" Nov 24 17:56:44 crc kubenswrapper[4777]: I1124 17:56:44.270534 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:56:44 crc kubenswrapper[4777]: I1124 17:56:44.271499 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:56:52 crc kubenswrapper[4777]: I1124 17:56:52.948401 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zmdsl_35059f02-39e8-40b0-8295-8230591cd996/control-plane-machine-set-operator/0.log" Nov 24 17:56:53 crc kubenswrapper[4777]: I1124 17:56:53.196228 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xtvn9_92efc884-606f-4fca-9410-4856b39a985d/kube-rbac-proxy/0.log" Nov 24 17:56:53 crc kubenswrapper[4777]: I1124 17:56:53.205418 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xtvn9_92efc884-606f-4fca-9410-4856b39a985d/machine-api-operator/0.log" Nov 24 17:57:06 crc kubenswrapper[4777]: I1124 17:57:06.183129 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-l75dg_0a75da4d-a94d-4639-ad88-0b415b87fe24/cert-manager-controller/0.log" Nov 24 17:57:06 crc kubenswrapper[4777]: I1124 17:57:06.318403 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-596z9_cee47616-4825-45f3-8232-4f46a66c954a/cert-manager-cainjector/0.log" Nov 24 17:57:06 crc kubenswrapper[4777]: I1124 17:57:06.394446 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-9f892_c32a0d11-9b78-4ce6-a4db-d9d07c9f3313/cert-manager-webhook/0.log" Nov 24 17:57:14 crc kubenswrapper[4777]: I1124 17:57:14.270591 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:57:14 crc kubenswrapper[4777]: I1124 17:57:14.271427 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:57:21 crc kubenswrapper[4777]: I1124 17:57:21.064166 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-p6sjq_2f2d18e5-97fa-47f6-9194-760ff54fdddc/nmstate-console-plugin/0.log" Nov 24 17:57:21 crc kubenswrapper[4777]: I1124 17:57:21.240107 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-wnlkc_6aa57fcf-4602-4a14-8f93-d9881aad4c8d/nmstate-handler/0.log" Nov 24 17:57:21 crc kubenswrapper[4777]: I1124 17:57:21.385222 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-m7wgh_2cb2400c-9df2-4e5f-95e9-3e8d22b58d3c/nmstate-metrics/0.log" Nov 24 17:57:21 crc kubenswrapper[4777]: I1124 17:57:21.390263 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-m7wgh_2cb2400c-9df2-4e5f-95e9-3e8d22b58d3c/kube-rbac-proxy/0.log" Nov 24 17:57:21 crc kubenswrapper[4777]: I1124 17:57:21.650514 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-8wtx5_3ffb9c2f-e94c-4fda-8c82-a34c045f9435/nmstate-operator/0.log" Nov 24 17:57:21 crc kubenswrapper[4777]: I1124 17:57:21.678962 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-z59z6_64c0b83e-a764-4494-b384-727c83c914a2/nmstate-webhook/0.log" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.591438 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zb9lv"] Nov 24 17:57:33 crc kubenswrapper[4777]: E1124 17:57:33.592307 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e418e82-f285-4adf-a6ac-e425aa66c9c8" containerName="extract-utilities" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.592319 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e418e82-f285-4adf-a6ac-e425aa66c9c8" containerName="extract-utilities" Nov 24 17:57:33 crc kubenswrapper[4777]: E1124 17:57:33.592344 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f65ef1-339d-486f-8fe1-4670be8a14e7" containerName="extract-utilities" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.592350 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f65ef1-339d-486f-8fe1-4670be8a14e7" containerName="extract-utilities" Nov 24 17:57:33 crc kubenswrapper[4777]: E1124 17:57:33.592363 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f65ef1-339d-486f-8fe1-4670be8a14e7" containerName="extract-content" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.592369 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f65ef1-339d-486f-8fe1-4670be8a14e7" containerName="extract-content" Nov 24 17:57:33 crc kubenswrapper[4777]: E1124 17:57:33.592377 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e418e82-f285-4adf-a6ac-e425aa66c9c8" containerName="registry-server" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.592382 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e418e82-f285-4adf-a6ac-e425aa66c9c8" containerName="registry-server" Nov 24 17:57:33 crc kubenswrapper[4777]: E1124 17:57:33.592400 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e418e82-f285-4adf-a6ac-e425aa66c9c8" containerName="extract-content" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.592406 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e418e82-f285-4adf-a6ac-e425aa66c9c8" containerName="extract-content" Nov 24 17:57:33 crc kubenswrapper[4777]: E1124 17:57:33.592416 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f65ef1-339d-486f-8fe1-4670be8a14e7" containerName="registry-server" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.592421 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f65ef1-339d-486f-8fe1-4670be8a14e7" containerName="registry-server" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.592611 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="91f65ef1-339d-486f-8fe1-4670be8a14e7" containerName="registry-server" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.592630 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e418e82-f285-4adf-a6ac-e425aa66c9c8" containerName="registry-server" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.594099 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.602603 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zb9lv"] Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.715468 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5941e247-cb8a-4b3e-81f8-09d0268673b2-catalog-content\") pod \"community-operators-zb9lv\" (UID: \"5941e247-cb8a-4b3e-81f8-09d0268673b2\") " pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.715607 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5941e247-cb8a-4b3e-81f8-09d0268673b2-utilities\") pod \"community-operators-zb9lv\" (UID: \"5941e247-cb8a-4b3e-81f8-09d0268673b2\") " pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.715651 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lxtr\" (UniqueName: \"kubernetes.io/projected/5941e247-cb8a-4b3e-81f8-09d0268673b2-kube-api-access-8lxtr\") pod \"community-operators-zb9lv\" (UID: \"5941e247-cb8a-4b3e-81f8-09d0268673b2\") " pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.817516 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lxtr\" (UniqueName: \"kubernetes.io/projected/5941e247-cb8a-4b3e-81f8-09d0268673b2-kube-api-access-8lxtr\") pod \"community-operators-zb9lv\" (UID: \"5941e247-cb8a-4b3e-81f8-09d0268673b2\") " pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.817659 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5941e247-cb8a-4b3e-81f8-09d0268673b2-catalog-content\") pod \"community-operators-zb9lv\" (UID: \"5941e247-cb8a-4b3e-81f8-09d0268673b2\") " pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.817795 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5941e247-cb8a-4b3e-81f8-09d0268673b2-utilities\") pod \"community-operators-zb9lv\" (UID: \"5941e247-cb8a-4b3e-81f8-09d0268673b2\") " pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.818359 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5941e247-cb8a-4b3e-81f8-09d0268673b2-utilities\") pod \"community-operators-zb9lv\" (UID: \"5941e247-cb8a-4b3e-81f8-09d0268673b2\") " pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.818520 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5941e247-cb8a-4b3e-81f8-09d0268673b2-catalog-content\") pod \"community-operators-zb9lv\" (UID: \"5941e247-cb8a-4b3e-81f8-09d0268673b2\") " pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.841847 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lxtr\" (UniqueName: \"kubernetes.io/projected/5941e247-cb8a-4b3e-81f8-09d0268673b2-kube-api-access-8lxtr\") pod \"community-operators-zb9lv\" (UID: \"5941e247-cb8a-4b3e-81f8-09d0268673b2\") " pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:33 crc kubenswrapper[4777]: I1124 17:57:33.917741 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:34 crc kubenswrapper[4777]: I1124 17:57:34.614853 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zb9lv"] Nov 24 17:57:35 crc kubenswrapper[4777]: I1124 17:57:35.580851 4777 generic.go:334] "Generic (PLEG): container finished" podID="5941e247-cb8a-4b3e-81f8-09d0268673b2" containerID="09bfb3c1b23c11d6e5e816f3190c0fd9193aceeeba7fbf030b64e3f6af600d24" exitCode=0 Nov 24 17:57:35 crc kubenswrapper[4777]: I1124 17:57:35.580937 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb9lv" event={"ID":"5941e247-cb8a-4b3e-81f8-09d0268673b2","Type":"ContainerDied","Data":"09bfb3c1b23c11d6e5e816f3190c0fd9193aceeeba7fbf030b64e3f6af600d24"} Nov 24 17:57:35 crc kubenswrapper[4777]: I1124 17:57:35.581430 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb9lv" event={"ID":"5941e247-cb8a-4b3e-81f8-09d0268673b2","Type":"ContainerStarted","Data":"fb1a348b52260bb9845c194d60201d4e40c110dfae91d9c16f39e83271258483"} Nov 24 17:57:36 crc kubenswrapper[4777]: I1124 17:57:36.979320 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6ffccfff7d-t8ltx_d1c8f636-6ede-4469-b270-63d3823d6507/manager/0.log" Nov 24 17:57:36 crc kubenswrapper[4777]: I1124 17:57:36.996218 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6ffccfff7d-t8ltx_d1c8f636-6ede-4469-b270-63d3823d6507/kube-rbac-proxy/0.log" Nov 24 17:57:37 crc kubenswrapper[4777]: I1124 17:57:37.616160 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb9lv" event={"ID":"5941e247-cb8a-4b3e-81f8-09d0268673b2","Type":"ContainerStarted","Data":"38ac8dbadd37bb756e79e4f1b8d8017f5cfc3a11f1e03c2aba05ff06e0dc7404"} Nov 24 17:57:39 crc kubenswrapper[4777]: I1124 17:57:39.635201 4777 generic.go:334] "Generic (PLEG): container finished" podID="5941e247-cb8a-4b3e-81f8-09d0268673b2" containerID="38ac8dbadd37bb756e79e4f1b8d8017f5cfc3a11f1e03c2aba05ff06e0dc7404" exitCode=0 Nov 24 17:57:39 crc kubenswrapper[4777]: I1124 17:57:39.635284 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb9lv" event={"ID":"5941e247-cb8a-4b3e-81f8-09d0268673b2","Type":"ContainerDied","Data":"38ac8dbadd37bb756e79e4f1b8d8017f5cfc3a11f1e03c2aba05ff06e0dc7404"} Nov 24 17:57:40 crc kubenswrapper[4777]: I1124 17:57:40.647040 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb9lv" event={"ID":"5941e247-cb8a-4b3e-81f8-09d0268673b2","Type":"ContainerStarted","Data":"a17c00b71c193501b5e203d3f91c55d598c92d0b7da6e91f4f530b314b6d3413"} Nov 24 17:57:40 crc kubenswrapper[4777]: I1124 17:57:40.671536 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zb9lv" podStartSLOduration=3.144496801 podStartE2EDuration="7.671517522s" podCreationTimestamp="2025-11-24 17:57:33 +0000 UTC" firstStartedPulling="2025-11-24 17:57:35.583134851 +0000 UTC m=+3413.742329900" lastFinishedPulling="2025-11-24 17:57:40.110155572 +0000 UTC m=+3418.269350621" observedRunningTime="2025-11-24 17:57:40.665391617 +0000 UTC m=+3418.824586726" watchObservedRunningTime="2025-11-24 17:57:40.671517522 +0000 UTC m=+3418.830712561" Nov 24 17:57:43 crc kubenswrapper[4777]: I1124 17:57:43.919744 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:43 crc kubenswrapper[4777]: I1124 17:57:43.920335 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:43 crc kubenswrapper[4777]: I1124 17:57:43.970474 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:44 crc kubenswrapper[4777]: I1124 17:57:44.270754 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:57:44 crc kubenswrapper[4777]: I1124 17:57:44.270818 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 17:57:44 crc kubenswrapper[4777]: I1124 17:57:44.270865 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 17:57:44 crc kubenswrapper[4777]: I1124 17:57:44.271681 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4232c66077cf2691260b3a2ead0239df3749a3cf336096399a2eb7ed33863917"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 17:57:44 crc kubenswrapper[4777]: I1124 17:57:44.271756 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://4232c66077cf2691260b3a2ead0239df3749a3cf336096399a2eb7ed33863917" gracePeriod=600 Nov 24 17:57:44 crc kubenswrapper[4777]: I1124 17:57:44.686222 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="4232c66077cf2691260b3a2ead0239df3749a3cf336096399a2eb7ed33863917" exitCode=0 Nov 24 17:57:44 crc kubenswrapper[4777]: I1124 17:57:44.686283 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"4232c66077cf2691260b3a2ead0239df3749a3cf336096399a2eb7ed33863917"} Nov 24 17:57:44 crc kubenswrapper[4777]: I1124 17:57:44.686857 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6"} Nov 24 17:57:44 crc kubenswrapper[4777]: I1124 17:57:44.686875 4777 scope.go:117] "RemoveContainer" containerID="48e23a0a5dcd45f320aad7fad9b07c8ad6594786d764b54c3068c26bc7699fd9" Nov 24 17:57:53 crc kubenswrapper[4777]: I1124 17:57:53.445421 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-z9z24_b4a78161-cf5f-4307-9adb-a05de51d620f/kube-rbac-proxy/0.log" Nov 24 17:57:53 crc kubenswrapper[4777]: I1124 17:57:53.473873 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-z9z24_b4a78161-cf5f-4307-9adb-a05de51d620f/controller/0.log" Nov 24 17:57:53 crc kubenswrapper[4777]: I1124 17:57:53.708119 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-frr-files/0.log" Nov 24 17:57:53 crc kubenswrapper[4777]: I1124 17:57:53.937808 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-reloader/0.log" Nov 24 17:57:53 crc kubenswrapper[4777]: I1124 17:57:53.943507 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-metrics/0.log" Nov 24 17:57:53 crc kubenswrapper[4777]: I1124 17:57:53.998165 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:54 crc kubenswrapper[4777]: I1124 17:57:54.002991 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-frr-files/0.log" Nov 24 17:57:54 crc kubenswrapper[4777]: I1124 17:57:54.028932 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-reloader/0.log" Nov 24 17:57:54 crc kubenswrapper[4777]: I1124 17:57:54.053497 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zb9lv"] Nov 24 17:57:54 crc kubenswrapper[4777]: I1124 17:57:54.273167 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-frr-files/0.log" Nov 24 17:57:54 crc kubenswrapper[4777]: I1124 17:57:54.331987 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-reloader/0.log" Nov 24 17:57:54 crc kubenswrapper[4777]: I1124 17:57:54.515289 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-metrics/0.log" Nov 24 17:57:54 crc kubenswrapper[4777]: I1124 17:57:54.571575 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-metrics/0.log" Nov 24 17:57:54 crc kubenswrapper[4777]: I1124 17:57:54.695242 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-frr-files/0.log" Nov 24 17:57:54 crc kubenswrapper[4777]: I1124 17:57:54.751214 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-reloader/0.log" Nov 24 17:57:54 crc kubenswrapper[4777]: I1124 17:57:54.793133 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-metrics/0.log" Nov 24 17:57:54 crc kubenswrapper[4777]: I1124 17:57:54.804104 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zb9lv" podUID="5941e247-cb8a-4b3e-81f8-09d0268673b2" containerName="registry-server" containerID="cri-o://a17c00b71c193501b5e203d3f91c55d598c92d0b7da6e91f4f530b314b6d3413" gracePeriod=2 Nov 24 17:57:54 crc kubenswrapper[4777]: I1124 17:57:54.813145 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/controller/0.log" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.047363 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/frr-metrics/0.log" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.079607 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/kube-rbac-proxy/0.log" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.117279 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/kube-rbac-proxy-frr/0.log" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.468427 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/reloader/0.log" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.473043 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-86vqd_2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a/frr-k8s-webhook-server/0.log" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.729874 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.788225 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5941e247-cb8a-4b3e-81f8-09d0268673b2-catalog-content\") pod \"5941e247-cb8a-4b3e-81f8-09d0268673b2\" (UID: \"5941e247-cb8a-4b3e-81f8-09d0268673b2\") " Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.788384 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lxtr\" (UniqueName: \"kubernetes.io/projected/5941e247-cb8a-4b3e-81f8-09d0268673b2-kube-api-access-8lxtr\") pod \"5941e247-cb8a-4b3e-81f8-09d0268673b2\" (UID: \"5941e247-cb8a-4b3e-81f8-09d0268673b2\") " Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.788421 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5941e247-cb8a-4b3e-81f8-09d0268673b2-utilities\") pod \"5941e247-cb8a-4b3e-81f8-09d0268673b2\" (UID: \"5941e247-cb8a-4b3e-81f8-09d0268673b2\") " Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.792604 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5941e247-cb8a-4b3e-81f8-09d0268673b2-utilities" (OuterVolumeSpecName: "utilities") pod "5941e247-cb8a-4b3e-81f8-09d0268673b2" (UID: "5941e247-cb8a-4b3e-81f8-09d0268673b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.803107 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5941e247-cb8a-4b3e-81f8-09d0268673b2-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.805469 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5941e247-cb8a-4b3e-81f8-09d0268673b2-kube-api-access-8lxtr" (OuterVolumeSpecName: "kube-api-access-8lxtr") pod "5941e247-cb8a-4b3e-81f8-09d0268673b2" (UID: "5941e247-cb8a-4b3e-81f8-09d0268673b2"). InnerVolumeSpecName "kube-api-access-8lxtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.817051 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7444c76584-h5nzb_364328a8-4ddb-4c3c-96bd-3a87ef271678/manager/0.log" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.837106 4777 generic.go:334] "Generic (PLEG): container finished" podID="5941e247-cb8a-4b3e-81f8-09d0268673b2" containerID="a17c00b71c193501b5e203d3f91c55d598c92d0b7da6e91f4f530b314b6d3413" exitCode=0 Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.837150 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb9lv" event={"ID":"5941e247-cb8a-4b3e-81f8-09d0268673b2","Type":"ContainerDied","Data":"a17c00b71c193501b5e203d3f91c55d598c92d0b7da6e91f4f530b314b6d3413"} Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.837177 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb9lv" event={"ID":"5941e247-cb8a-4b3e-81f8-09d0268673b2","Type":"ContainerDied","Data":"fb1a348b52260bb9845c194d60201d4e40c110dfae91d9c16f39e83271258483"} Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.837194 4777 scope.go:117] "RemoveContainer" containerID="a17c00b71c193501b5e203d3f91c55d598c92d0b7da6e91f4f530b314b6d3413" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.837322 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zb9lv" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.849475 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5941e247-cb8a-4b3e-81f8-09d0268673b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5941e247-cb8a-4b3e-81f8-09d0268673b2" (UID: "5941e247-cb8a-4b3e-81f8-09d0268673b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.883346 4777 scope.go:117] "RemoveContainer" containerID="38ac8dbadd37bb756e79e4f1b8d8017f5cfc3a11f1e03c2aba05ff06e0dc7404" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.904618 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5941e247-cb8a-4b3e-81f8-09d0268673b2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.904655 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lxtr\" (UniqueName: \"kubernetes.io/projected/5941e247-cb8a-4b3e-81f8-09d0268673b2-kube-api-access-8lxtr\") on node \"crc\" DevicePath \"\"" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.940141 4777 scope.go:117] "RemoveContainer" containerID="09bfb3c1b23c11d6e5e816f3190c0fd9193aceeeba7fbf030b64e3f6af600d24" Nov 24 17:57:55 crc kubenswrapper[4777]: I1124 17:57:55.962420 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7578594d7b-kn9dd_61277f6c-5e55-4fdb-a482-e9b3a91edd7f/webhook-server/0.log" Nov 24 17:57:56 crc kubenswrapper[4777]: I1124 17:57:56.002618 4777 scope.go:117] "RemoveContainer" containerID="a17c00b71c193501b5e203d3f91c55d598c92d0b7da6e91f4f530b314b6d3413" Nov 24 17:57:56 crc kubenswrapper[4777]: E1124 17:57:56.004122 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a17c00b71c193501b5e203d3f91c55d598c92d0b7da6e91f4f530b314b6d3413\": container with ID starting with a17c00b71c193501b5e203d3f91c55d598c92d0b7da6e91f4f530b314b6d3413 not found: ID does not exist" containerID="a17c00b71c193501b5e203d3f91c55d598c92d0b7da6e91f4f530b314b6d3413" Nov 24 17:57:56 crc kubenswrapper[4777]: I1124 17:57:56.004175 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a17c00b71c193501b5e203d3f91c55d598c92d0b7da6e91f4f530b314b6d3413"} err="failed to get container status \"a17c00b71c193501b5e203d3f91c55d598c92d0b7da6e91f4f530b314b6d3413\": rpc error: code = NotFound desc = could not find container \"a17c00b71c193501b5e203d3f91c55d598c92d0b7da6e91f4f530b314b6d3413\": container with ID starting with a17c00b71c193501b5e203d3f91c55d598c92d0b7da6e91f4f530b314b6d3413 not found: ID does not exist" Nov 24 17:57:56 crc kubenswrapper[4777]: I1124 17:57:56.004201 4777 scope.go:117] "RemoveContainer" containerID="38ac8dbadd37bb756e79e4f1b8d8017f5cfc3a11f1e03c2aba05ff06e0dc7404" Nov 24 17:57:56 crc kubenswrapper[4777]: E1124 17:57:56.005205 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38ac8dbadd37bb756e79e4f1b8d8017f5cfc3a11f1e03c2aba05ff06e0dc7404\": container with ID starting with 38ac8dbadd37bb756e79e4f1b8d8017f5cfc3a11f1e03c2aba05ff06e0dc7404 not found: ID does not exist" containerID="38ac8dbadd37bb756e79e4f1b8d8017f5cfc3a11f1e03c2aba05ff06e0dc7404" Nov 24 17:57:56 crc kubenswrapper[4777]: I1124 17:57:56.005238 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38ac8dbadd37bb756e79e4f1b8d8017f5cfc3a11f1e03c2aba05ff06e0dc7404"} err="failed to get container status \"38ac8dbadd37bb756e79e4f1b8d8017f5cfc3a11f1e03c2aba05ff06e0dc7404\": rpc error: code = NotFound desc = could not find container \"38ac8dbadd37bb756e79e4f1b8d8017f5cfc3a11f1e03c2aba05ff06e0dc7404\": container with ID starting with 38ac8dbadd37bb756e79e4f1b8d8017f5cfc3a11f1e03c2aba05ff06e0dc7404 not found: ID does not exist" Nov 24 17:57:56 crc kubenswrapper[4777]: I1124 17:57:56.005259 4777 scope.go:117] "RemoveContainer" containerID="09bfb3c1b23c11d6e5e816f3190c0fd9193aceeeba7fbf030b64e3f6af600d24" Nov 24 17:57:56 crc kubenswrapper[4777]: E1124 17:57:56.006705 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09bfb3c1b23c11d6e5e816f3190c0fd9193aceeeba7fbf030b64e3f6af600d24\": container with ID starting with 09bfb3c1b23c11d6e5e816f3190c0fd9193aceeeba7fbf030b64e3f6af600d24 not found: ID does not exist" containerID="09bfb3c1b23c11d6e5e816f3190c0fd9193aceeeba7fbf030b64e3f6af600d24" Nov 24 17:57:56 crc kubenswrapper[4777]: I1124 17:57:56.017100 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09bfb3c1b23c11d6e5e816f3190c0fd9193aceeeba7fbf030b64e3f6af600d24"} err="failed to get container status \"09bfb3c1b23c11d6e5e816f3190c0fd9193aceeeba7fbf030b64e3f6af600d24\": rpc error: code = NotFound desc = could not find container \"09bfb3c1b23c11d6e5e816f3190c0fd9193aceeeba7fbf030b64e3f6af600d24\": container with ID starting with 09bfb3c1b23c11d6e5e816f3190c0fd9193aceeeba7fbf030b64e3f6af600d24 not found: ID does not exist" Nov 24 17:57:56 crc kubenswrapper[4777]: I1124 17:57:56.137710 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-pz9ts_5ae956fa-5a45-4248-b300-d802aa171d52/kube-rbac-proxy/0.log" Nov 24 17:57:56 crc kubenswrapper[4777]: I1124 17:57:56.181597 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zb9lv"] Nov 24 17:57:56 crc kubenswrapper[4777]: I1124 17:57:56.198168 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zb9lv"] Nov 24 17:57:56 crc kubenswrapper[4777]: I1124 17:57:56.313544 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/frr/0.log" Nov 24 17:57:56 crc kubenswrapper[4777]: I1124 17:57:56.547383 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-pz9ts_5ae956fa-5a45-4248-b300-d802aa171d52/speaker/0.log" Nov 24 17:57:57 crc kubenswrapper[4777]: I1124 17:57:57.262348 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5941e247-cb8a-4b3e-81f8-09d0268673b2" path="/var/lib/kubelet/pods/5941e247-cb8a-4b3e-81f8-09d0268673b2/volumes" Nov 24 17:58:10 crc kubenswrapper[4777]: I1124 17:58:10.778448 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c_1ffbfe00-f54c-446d-bc90-d18f24f51f82/util/0.log" Nov 24 17:58:10 crc kubenswrapper[4777]: I1124 17:58:10.995253 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c_1ffbfe00-f54c-446d-bc90-d18f24f51f82/util/0.log" Nov 24 17:58:11 crc kubenswrapper[4777]: I1124 17:58:11.007864 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c_1ffbfe00-f54c-446d-bc90-d18f24f51f82/pull/0.log" Nov 24 17:58:11 crc kubenswrapper[4777]: I1124 17:58:11.017007 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c_1ffbfe00-f54c-446d-bc90-d18f24f51f82/pull/0.log" Nov 24 17:58:11 crc kubenswrapper[4777]: I1124 17:58:11.157981 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c_1ffbfe00-f54c-446d-bc90-d18f24f51f82/util/0.log" Nov 24 17:58:11 crc kubenswrapper[4777]: I1124 17:58:11.201454 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c_1ffbfe00-f54c-446d-bc90-d18f24f51f82/extract/0.log" Nov 24 17:58:11 crc kubenswrapper[4777]: I1124 17:58:11.225113 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c_1ffbfe00-f54c-446d-bc90-d18f24f51f82/pull/0.log" Nov 24 17:58:11 crc kubenswrapper[4777]: I1124 17:58:11.392399 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd_edece1c5-1d2f-4a61-9591-7313cd41d58d/util/0.log" Nov 24 17:58:11 crc kubenswrapper[4777]: I1124 17:58:11.519950 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd_edece1c5-1d2f-4a61-9591-7313cd41d58d/pull/0.log" Nov 24 17:58:11 crc kubenswrapper[4777]: I1124 17:58:11.589890 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd_edece1c5-1d2f-4a61-9591-7313cd41d58d/util/0.log" Nov 24 17:58:11 crc kubenswrapper[4777]: I1124 17:58:11.607033 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd_edece1c5-1d2f-4a61-9591-7313cd41d58d/pull/0.log" Nov 24 17:58:11 crc kubenswrapper[4777]: I1124 17:58:11.765652 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd_edece1c5-1d2f-4a61-9591-7313cd41d58d/util/0.log" Nov 24 17:58:11 crc kubenswrapper[4777]: I1124 17:58:11.788983 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd_edece1c5-1d2f-4a61-9591-7313cd41d58d/extract/0.log" Nov 24 17:58:11 crc kubenswrapper[4777]: I1124 17:58:11.791170 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd_edece1c5-1d2f-4a61-9591-7313cd41d58d/pull/0.log" Nov 24 17:58:11 crc kubenswrapper[4777]: I1124 17:58:11.951425 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww_213b5cf0-7684-4331-b5be-28fd188e0db3/util/0.log" Nov 24 17:58:12 crc kubenswrapper[4777]: I1124 17:58:12.115131 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww_213b5cf0-7684-4331-b5be-28fd188e0db3/pull/0.log" Nov 24 17:58:12 crc kubenswrapper[4777]: I1124 17:58:12.119403 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww_213b5cf0-7684-4331-b5be-28fd188e0db3/pull/0.log" Nov 24 17:58:12 crc kubenswrapper[4777]: I1124 17:58:12.142814 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww_213b5cf0-7684-4331-b5be-28fd188e0db3/util/0.log" Nov 24 17:58:12 crc kubenswrapper[4777]: I1124 17:58:12.347955 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww_213b5cf0-7684-4331-b5be-28fd188e0db3/pull/0.log" Nov 24 17:58:12 crc kubenswrapper[4777]: I1124 17:58:12.393898 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww_213b5cf0-7684-4331-b5be-28fd188e0db3/extract/0.log" Nov 24 17:58:12 crc kubenswrapper[4777]: I1124 17:58:12.418305 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww_213b5cf0-7684-4331-b5be-28fd188e0db3/util/0.log" Nov 24 17:58:12 crc kubenswrapper[4777]: I1124 17:58:12.560802 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9_9194629c-e5c7-4ef7-8b08-67ae8d58a1a2/util/0.log" Nov 24 17:58:12 crc kubenswrapper[4777]: I1124 17:58:12.770089 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9_9194629c-e5c7-4ef7-8b08-67ae8d58a1a2/pull/0.log" Nov 24 17:58:12 crc kubenswrapper[4777]: I1124 17:58:12.791113 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9_9194629c-e5c7-4ef7-8b08-67ae8d58a1a2/util/0.log" Nov 24 17:58:12 crc kubenswrapper[4777]: I1124 17:58:12.813791 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9_9194629c-e5c7-4ef7-8b08-67ae8d58a1a2/pull/0.log" Nov 24 17:58:12 crc kubenswrapper[4777]: I1124 17:58:12.971506 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9_9194629c-e5c7-4ef7-8b08-67ae8d58a1a2/util/0.log" Nov 24 17:58:12 crc kubenswrapper[4777]: I1124 17:58:12.975803 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9_9194629c-e5c7-4ef7-8b08-67ae8d58a1a2/extract/0.log" Nov 24 17:58:12 crc kubenswrapper[4777]: I1124 17:58:12.986181 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9_9194629c-e5c7-4ef7-8b08-67ae8d58a1a2/pull/0.log" Nov 24 17:58:13 crc kubenswrapper[4777]: I1124 17:58:13.160550 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5m4fr_6aa2fa6f-bd34-4154-b666-455e7fea71ae/extract-utilities/0.log" Nov 24 17:58:13 crc kubenswrapper[4777]: I1124 17:58:13.405294 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5m4fr_6aa2fa6f-bd34-4154-b666-455e7fea71ae/extract-content/0.log" Nov 24 17:58:13 crc kubenswrapper[4777]: I1124 17:58:13.412675 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5m4fr_6aa2fa6f-bd34-4154-b666-455e7fea71ae/extract-content/0.log" Nov 24 17:58:13 crc kubenswrapper[4777]: I1124 17:58:13.420896 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5m4fr_6aa2fa6f-bd34-4154-b666-455e7fea71ae/extract-utilities/0.log" Nov 24 17:58:13 crc kubenswrapper[4777]: I1124 17:58:13.656077 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5m4fr_6aa2fa6f-bd34-4154-b666-455e7fea71ae/extract-content/0.log" Nov 24 17:58:13 crc kubenswrapper[4777]: I1124 17:58:13.675564 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5m4fr_6aa2fa6f-bd34-4154-b666-455e7fea71ae/extract-utilities/0.log" Nov 24 17:58:13 crc kubenswrapper[4777]: I1124 17:58:13.965693 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h54gs_30623966-678a-4124-bc99-02d4ccfacb85/extract-utilities/0.log" Nov 24 17:58:14 crc kubenswrapper[4777]: I1124 17:58:14.143472 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h54gs_30623966-678a-4124-bc99-02d4ccfacb85/extract-content/0.log" Nov 24 17:58:14 crc kubenswrapper[4777]: I1124 17:58:14.200407 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h54gs_30623966-678a-4124-bc99-02d4ccfacb85/extract-utilities/0.log" Nov 24 17:58:14 crc kubenswrapper[4777]: I1124 17:58:14.211644 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5m4fr_6aa2fa6f-bd34-4154-b666-455e7fea71ae/registry-server/0.log" Nov 24 17:58:14 crc kubenswrapper[4777]: I1124 17:58:14.259618 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h54gs_30623966-678a-4124-bc99-02d4ccfacb85/extract-content/0.log" Nov 24 17:58:14 crc kubenswrapper[4777]: I1124 17:58:14.459823 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h54gs_30623966-678a-4124-bc99-02d4ccfacb85/extract-utilities/0.log" Nov 24 17:58:14 crc kubenswrapper[4777]: I1124 17:58:14.476399 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h54gs_30623966-678a-4124-bc99-02d4ccfacb85/extract-content/0.log" Nov 24 17:58:14 crc kubenswrapper[4777]: I1124 17:58:14.653020 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4_539a904e-3754-42f1-95ec-13140964520d/util/0.log" Nov 24 17:58:14 crc kubenswrapper[4777]: I1124 17:58:14.826732 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4_539a904e-3754-42f1-95ec-13140964520d/pull/0.log" Nov 24 17:58:14 crc kubenswrapper[4777]: I1124 17:58:14.836026 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4_539a904e-3754-42f1-95ec-13140964520d/pull/0.log" Nov 24 17:58:14 crc kubenswrapper[4777]: I1124 17:58:14.842790 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4_539a904e-3754-42f1-95ec-13140964520d/util/0.log" Nov 24 17:58:15 crc kubenswrapper[4777]: I1124 17:58:15.030520 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h54gs_30623966-678a-4124-bc99-02d4ccfacb85/registry-server/0.log" Nov 24 17:58:15 crc kubenswrapper[4777]: I1124 17:58:15.084021 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4_539a904e-3754-42f1-95ec-13140964520d/pull/0.log" Nov 24 17:58:15 crc kubenswrapper[4777]: I1124 17:58:15.089670 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4_539a904e-3754-42f1-95ec-13140964520d/extract/0.log" Nov 24 17:58:15 crc kubenswrapper[4777]: I1124 17:58:15.117520 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4_539a904e-3754-42f1-95ec-13140964520d/util/0.log" Nov 24 17:58:15 crc kubenswrapper[4777]: I1124 17:58:15.238094 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-5dsxp_3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da/marketplace-operator/0.log" Nov 24 17:58:15 crc kubenswrapper[4777]: I1124 17:58:15.277157 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hwftb_27733c5c-8536-4f12-8aa0-e5a9d16e8f88/extract-utilities/0.log" Nov 24 17:58:15 crc kubenswrapper[4777]: I1124 17:58:15.529450 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hwftb_27733c5c-8536-4f12-8aa0-e5a9d16e8f88/extract-content/0.log" Nov 24 17:58:15 crc kubenswrapper[4777]: I1124 17:58:15.551266 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hwftb_27733c5c-8536-4f12-8aa0-e5a9d16e8f88/extract-utilities/0.log" Nov 24 17:58:15 crc kubenswrapper[4777]: I1124 17:58:15.556518 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hwftb_27733c5c-8536-4f12-8aa0-e5a9d16e8f88/extract-content/0.log" Nov 24 17:58:15 crc kubenswrapper[4777]: I1124 17:58:15.776727 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hwftb_27733c5c-8536-4f12-8aa0-e5a9d16e8f88/extract-content/0.log" Nov 24 17:58:15 crc kubenswrapper[4777]: I1124 17:58:15.776865 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hwftb_27733c5c-8536-4f12-8aa0-e5a9d16e8f88/extract-utilities/0.log" Nov 24 17:58:15 crc kubenswrapper[4777]: I1124 17:58:15.864199 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g7f5l_1af51c54-3a40-406e-8743-32b0c04d018f/extract-utilities/0.log" Nov 24 17:58:15 crc kubenswrapper[4777]: I1124 17:58:15.904770 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hwftb_27733c5c-8536-4f12-8aa0-e5a9d16e8f88/registry-server/0.log" Nov 24 17:58:16 crc kubenswrapper[4777]: I1124 17:58:16.053110 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g7f5l_1af51c54-3a40-406e-8743-32b0c04d018f/extract-content/0.log" Nov 24 17:58:16 crc kubenswrapper[4777]: I1124 17:58:16.084354 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g7f5l_1af51c54-3a40-406e-8743-32b0c04d018f/extract-utilities/0.log" Nov 24 17:58:16 crc kubenswrapper[4777]: I1124 17:58:16.084394 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g7f5l_1af51c54-3a40-406e-8743-32b0c04d018f/extract-content/0.log" Nov 24 17:58:16 crc kubenswrapper[4777]: I1124 17:58:16.264951 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g7f5l_1af51c54-3a40-406e-8743-32b0c04d018f/extract-utilities/0.log" Nov 24 17:58:16 crc kubenswrapper[4777]: I1124 17:58:16.311285 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g7f5l_1af51c54-3a40-406e-8743-32b0c04d018f/extract-content/0.log" Nov 24 17:58:16 crc kubenswrapper[4777]: I1124 17:58:16.837670 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g7f5l_1af51c54-3a40-406e-8743-32b0c04d018f/registry-server/0.log" Nov 24 17:58:29 crc kubenswrapper[4777]: I1124 17:58:29.137960 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-wsbfx_b08a2563-7e34-40f7-b9b0-052824422cff/prometheus-operator/0.log" Nov 24 17:58:29 crc kubenswrapper[4777]: I1124 17:58:29.317512 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_dd6addca-3fb6-4204-ae4e-2ed8d1326699/prometheus-operator-admission-webhook/0.log" Nov 24 17:58:29 crc kubenswrapper[4777]: I1124 17:58:29.355987 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_67418bcd-f0ba-4526-afc1-8420cb642ccf/prometheus-operator-admission-webhook/0.log" Nov 24 17:58:29 crc kubenswrapper[4777]: I1124 17:58:29.498191 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-hdxt8_986d062f-3210-471c-a2db-07e080b5d449/operator/0.log" Nov 24 17:58:29 crc kubenswrapper[4777]: I1124 17:58:29.519791 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-22cbh_c0fe98e7-9ad1-426e-a17a-ad471a7f73af/perses-operator/0.log" Nov 24 17:58:42 crc kubenswrapper[4777]: I1124 17:58:42.298601 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6ffccfff7d-t8ltx_d1c8f636-6ede-4469-b270-63d3823d6507/manager/0.log" Nov 24 17:58:42 crc kubenswrapper[4777]: I1124 17:58:42.310625 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6ffccfff7d-t8ltx_d1c8f636-6ede-4469-b270-63d3823d6507/kube-rbac-proxy/0.log" Nov 24 17:59:44 crc kubenswrapper[4777]: I1124 17:59:44.270063 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 17:59:44 crc kubenswrapper[4777]: I1124 17:59:44.271680 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.188006 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf"] Nov 24 18:00:00 crc kubenswrapper[4777]: E1124 18:00:00.189116 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5941e247-cb8a-4b3e-81f8-09d0268673b2" containerName="extract-utilities" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.189132 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="5941e247-cb8a-4b3e-81f8-09d0268673b2" containerName="extract-utilities" Nov 24 18:00:00 crc kubenswrapper[4777]: E1124 18:00:00.189180 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5941e247-cb8a-4b3e-81f8-09d0268673b2" containerName="registry-server" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.189190 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="5941e247-cb8a-4b3e-81f8-09d0268673b2" containerName="registry-server" Nov 24 18:00:00 crc kubenswrapper[4777]: E1124 18:00:00.189215 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5941e247-cb8a-4b3e-81f8-09d0268673b2" containerName="extract-content" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.189222 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="5941e247-cb8a-4b3e-81f8-09d0268673b2" containerName="extract-content" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.189458 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="5941e247-cb8a-4b3e-81f8-09d0268673b2" containerName="registry-server" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.190413 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.193314 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.193560 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.213003 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf"] Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.235212 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5cff5213-efb6-4d07-ae75-e9eb38583771-secret-volume\") pod \"collect-profiles-29400120-4tgjf\" (UID: \"5cff5213-efb6-4d07-ae75-e9eb38583771\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.235438 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5cff5213-efb6-4d07-ae75-e9eb38583771-config-volume\") pod \"collect-profiles-29400120-4tgjf\" (UID: \"5cff5213-efb6-4d07-ae75-e9eb38583771\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.235472 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb6fl\" (UniqueName: \"kubernetes.io/projected/5cff5213-efb6-4d07-ae75-e9eb38583771-kube-api-access-gb6fl\") pod \"collect-profiles-29400120-4tgjf\" (UID: \"5cff5213-efb6-4d07-ae75-e9eb38583771\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.336786 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5cff5213-efb6-4d07-ae75-e9eb38583771-config-volume\") pod \"collect-profiles-29400120-4tgjf\" (UID: \"5cff5213-efb6-4d07-ae75-e9eb38583771\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.336846 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb6fl\" (UniqueName: \"kubernetes.io/projected/5cff5213-efb6-4d07-ae75-e9eb38583771-kube-api-access-gb6fl\") pod \"collect-profiles-29400120-4tgjf\" (UID: \"5cff5213-efb6-4d07-ae75-e9eb38583771\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.337022 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5cff5213-efb6-4d07-ae75-e9eb38583771-secret-volume\") pod \"collect-profiles-29400120-4tgjf\" (UID: \"5cff5213-efb6-4d07-ae75-e9eb38583771\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.337772 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5cff5213-efb6-4d07-ae75-e9eb38583771-config-volume\") pod \"collect-profiles-29400120-4tgjf\" (UID: \"5cff5213-efb6-4d07-ae75-e9eb38583771\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.344437 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5cff5213-efb6-4d07-ae75-e9eb38583771-secret-volume\") pod \"collect-profiles-29400120-4tgjf\" (UID: \"5cff5213-efb6-4d07-ae75-e9eb38583771\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.352646 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb6fl\" (UniqueName: \"kubernetes.io/projected/5cff5213-efb6-4d07-ae75-e9eb38583771-kube-api-access-gb6fl\") pod \"collect-profiles-29400120-4tgjf\" (UID: \"5cff5213-efb6-4d07-ae75-e9eb38583771\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" Nov 24 18:00:00 crc kubenswrapper[4777]: I1124 18:00:00.534193 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" Nov 24 18:00:01 crc kubenswrapper[4777]: I1124 18:00:01.085362 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf"] Nov 24 18:00:01 crc kubenswrapper[4777]: I1124 18:00:01.125870 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" event={"ID":"5cff5213-efb6-4d07-ae75-e9eb38583771","Type":"ContainerStarted","Data":"57dc2d38dfce520fd49898359be769eab2dacd631d84ea7dee6bdb07065b2175"} Nov 24 18:00:02 crc kubenswrapper[4777]: I1124 18:00:02.140322 4777 generic.go:334] "Generic (PLEG): container finished" podID="5cff5213-efb6-4d07-ae75-e9eb38583771" containerID="74006a0eeb779ac6cbfe240d2b56e91ce1e6ec190d5201c768e7aabb8abdc392" exitCode=0 Nov 24 18:00:02 crc kubenswrapper[4777]: I1124 18:00:02.140422 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" event={"ID":"5cff5213-efb6-4d07-ae75-e9eb38583771","Type":"ContainerDied","Data":"74006a0eeb779ac6cbfe240d2b56e91ce1e6ec190d5201c768e7aabb8abdc392"} Nov 24 18:00:03 crc kubenswrapper[4777]: I1124 18:00:03.792485 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" Nov 24 18:00:03 crc kubenswrapper[4777]: I1124 18:00:03.929578 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5cff5213-efb6-4d07-ae75-e9eb38583771-config-volume\") pod \"5cff5213-efb6-4d07-ae75-e9eb38583771\" (UID: \"5cff5213-efb6-4d07-ae75-e9eb38583771\") " Nov 24 18:00:03 crc kubenswrapper[4777]: I1124 18:00:03.929693 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb6fl\" (UniqueName: \"kubernetes.io/projected/5cff5213-efb6-4d07-ae75-e9eb38583771-kube-api-access-gb6fl\") pod \"5cff5213-efb6-4d07-ae75-e9eb38583771\" (UID: \"5cff5213-efb6-4d07-ae75-e9eb38583771\") " Nov 24 18:00:03 crc kubenswrapper[4777]: I1124 18:00:03.929859 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5cff5213-efb6-4d07-ae75-e9eb38583771-secret-volume\") pod \"5cff5213-efb6-4d07-ae75-e9eb38583771\" (UID: \"5cff5213-efb6-4d07-ae75-e9eb38583771\") " Nov 24 18:00:03 crc kubenswrapper[4777]: I1124 18:00:03.931204 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cff5213-efb6-4d07-ae75-e9eb38583771-config-volume" (OuterVolumeSpecName: "config-volume") pod "5cff5213-efb6-4d07-ae75-e9eb38583771" (UID: "5cff5213-efb6-4d07-ae75-e9eb38583771"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 18:00:03 crc kubenswrapper[4777]: I1124 18:00:03.942688 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cff5213-efb6-4d07-ae75-e9eb38583771-kube-api-access-gb6fl" (OuterVolumeSpecName: "kube-api-access-gb6fl") pod "5cff5213-efb6-4d07-ae75-e9eb38583771" (UID: "5cff5213-efb6-4d07-ae75-e9eb38583771"). InnerVolumeSpecName "kube-api-access-gb6fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:00:03 crc kubenswrapper[4777]: I1124 18:00:03.942901 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cff5213-efb6-4d07-ae75-e9eb38583771-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5cff5213-efb6-4d07-ae75-e9eb38583771" (UID: "5cff5213-efb6-4d07-ae75-e9eb38583771"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:00:04 crc kubenswrapper[4777]: I1124 18:00:04.033797 4777 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5cff5213-efb6-4d07-ae75-e9eb38583771-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 18:00:04 crc kubenswrapper[4777]: I1124 18:00:04.033845 4777 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5cff5213-efb6-4d07-ae75-e9eb38583771-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 18:00:04 crc kubenswrapper[4777]: I1124 18:00:04.033858 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb6fl\" (UniqueName: \"kubernetes.io/projected/5cff5213-efb6-4d07-ae75-e9eb38583771-kube-api-access-gb6fl\") on node \"crc\" DevicePath \"\"" Nov 24 18:00:04 crc kubenswrapper[4777]: I1124 18:00:04.164823 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" event={"ID":"5cff5213-efb6-4d07-ae75-e9eb38583771","Type":"ContainerDied","Data":"57dc2d38dfce520fd49898359be769eab2dacd631d84ea7dee6bdb07065b2175"} Nov 24 18:00:04 crc kubenswrapper[4777]: I1124 18:00:04.164864 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57dc2d38dfce520fd49898359be769eab2dacd631d84ea7dee6bdb07065b2175" Nov 24 18:00:04 crc kubenswrapper[4777]: I1124 18:00:04.164925 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400120-4tgjf" Nov 24 18:00:04 crc kubenswrapper[4777]: I1124 18:00:04.861283 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb"] Nov 24 18:00:04 crc kubenswrapper[4777]: I1124 18:00:04.872851 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400075-rppzb"] Nov 24 18:00:05 crc kubenswrapper[4777]: I1124 18:00:05.259663 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="631cfc31-4721-408b-860c-e7675ed55892" path="/var/lib/kubelet/pods/631cfc31-4721-408b-860c-e7675ed55892/volumes" Nov 24 18:00:14 crc kubenswrapper[4777]: I1124 18:00:14.270272 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:00:14 crc kubenswrapper[4777]: I1124 18:00:14.271822 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:00:29 crc kubenswrapper[4777]: I1124 18:00:29.521637 4777 generic.go:334] "Generic (PLEG): container finished" podID="b575ae4b-aac8-4a05-a096-a85cae8f1718" containerID="3b073f64a3f303b2e807761c1532e2ea625276750f3196db0dff67b43d5828d2" exitCode=0 Nov 24 18:00:29 crc kubenswrapper[4777]: I1124 18:00:29.521828 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gf5c/must-gather-krqqk" event={"ID":"b575ae4b-aac8-4a05-a096-a85cae8f1718","Type":"ContainerDied","Data":"3b073f64a3f303b2e807761c1532e2ea625276750f3196db0dff67b43d5828d2"} Nov 24 18:00:29 crc kubenswrapper[4777]: I1124 18:00:29.522929 4777 scope.go:117] "RemoveContainer" containerID="3b073f64a3f303b2e807761c1532e2ea625276750f3196db0dff67b43d5828d2" Nov 24 18:00:29 crc kubenswrapper[4777]: I1124 18:00:29.605586 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4gf5c_must-gather-krqqk_b575ae4b-aac8-4a05-a096-a85cae8f1718/gather/0.log" Nov 24 18:00:38 crc kubenswrapper[4777]: I1124 18:00:38.145883 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4gf5c/must-gather-krqqk"] Nov 24 18:00:38 crc kubenswrapper[4777]: I1124 18:00:38.146610 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-4gf5c/must-gather-krqqk" podUID="b575ae4b-aac8-4a05-a096-a85cae8f1718" containerName="copy" containerID="cri-o://aa484d797efb71d3a21014ea068c0452dfd7ff17f9329a0b23417c7d9c675207" gracePeriod=2 Nov 24 18:00:38 crc kubenswrapper[4777]: I1124 18:00:38.158427 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4gf5c/must-gather-krqqk"] Nov 24 18:00:38 crc kubenswrapper[4777]: I1124 18:00:38.610009 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4gf5c_must-gather-krqqk_b575ae4b-aac8-4a05-a096-a85cae8f1718/copy/0.log" Nov 24 18:00:38 crc kubenswrapper[4777]: I1124 18:00:38.610820 4777 generic.go:334] "Generic (PLEG): container finished" podID="b575ae4b-aac8-4a05-a096-a85cae8f1718" containerID="aa484d797efb71d3a21014ea068c0452dfd7ff17f9329a0b23417c7d9c675207" exitCode=143 Nov 24 18:00:38 crc kubenswrapper[4777]: I1124 18:00:38.910002 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4gf5c_must-gather-krqqk_b575ae4b-aac8-4a05-a096-a85cae8f1718/copy/0.log" Nov 24 18:00:38 crc kubenswrapper[4777]: I1124 18:00:38.910440 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/must-gather-krqqk" Nov 24 18:00:39 crc kubenswrapper[4777]: I1124 18:00:39.016849 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b575ae4b-aac8-4a05-a096-a85cae8f1718-must-gather-output\") pod \"b575ae4b-aac8-4a05-a096-a85cae8f1718\" (UID: \"b575ae4b-aac8-4a05-a096-a85cae8f1718\") " Nov 24 18:00:39 crc kubenswrapper[4777]: I1124 18:00:39.017246 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cswr\" (UniqueName: \"kubernetes.io/projected/b575ae4b-aac8-4a05-a096-a85cae8f1718-kube-api-access-6cswr\") pod \"b575ae4b-aac8-4a05-a096-a85cae8f1718\" (UID: \"b575ae4b-aac8-4a05-a096-a85cae8f1718\") " Nov 24 18:00:39 crc kubenswrapper[4777]: I1124 18:00:39.026124 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b575ae4b-aac8-4a05-a096-a85cae8f1718-kube-api-access-6cswr" (OuterVolumeSpecName: "kube-api-access-6cswr") pod "b575ae4b-aac8-4a05-a096-a85cae8f1718" (UID: "b575ae4b-aac8-4a05-a096-a85cae8f1718"). InnerVolumeSpecName "kube-api-access-6cswr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:00:39 crc kubenswrapper[4777]: I1124 18:00:39.119364 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cswr\" (UniqueName: \"kubernetes.io/projected/b575ae4b-aac8-4a05-a096-a85cae8f1718-kube-api-access-6cswr\") on node \"crc\" DevicePath \"\"" Nov 24 18:00:39 crc kubenswrapper[4777]: I1124 18:00:39.172365 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b575ae4b-aac8-4a05-a096-a85cae8f1718-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b575ae4b-aac8-4a05-a096-a85cae8f1718" (UID: "b575ae4b-aac8-4a05-a096-a85cae8f1718"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:00:39 crc kubenswrapper[4777]: I1124 18:00:39.221696 4777 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b575ae4b-aac8-4a05-a096-a85cae8f1718-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 18:00:39 crc kubenswrapper[4777]: I1124 18:00:39.259067 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b575ae4b-aac8-4a05-a096-a85cae8f1718" path="/var/lib/kubelet/pods/b575ae4b-aac8-4a05-a096-a85cae8f1718/volumes" Nov 24 18:00:39 crc kubenswrapper[4777]: I1124 18:00:39.622749 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4gf5c_must-gather-krqqk_b575ae4b-aac8-4a05-a096-a85cae8f1718/copy/0.log" Nov 24 18:00:39 crc kubenswrapper[4777]: I1124 18:00:39.623154 4777 scope.go:117] "RemoveContainer" containerID="aa484d797efb71d3a21014ea068c0452dfd7ff17f9329a0b23417c7d9c675207" Nov 24 18:00:39 crc kubenswrapper[4777]: I1124 18:00:39.623247 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gf5c/must-gather-krqqk" Nov 24 18:00:39 crc kubenswrapper[4777]: I1124 18:00:39.642504 4777 scope.go:117] "RemoveContainer" containerID="3b073f64a3f303b2e807761c1532e2ea625276750f3196db0dff67b43d5828d2" Nov 24 18:00:44 crc kubenswrapper[4777]: I1124 18:00:44.269890 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:00:44 crc kubenswrapper[4777]: I1124 18:00:44.270729 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:00:44 crc kubenswrapper[4777]: I1124 18:00:44.270792 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 18:00:44 crc kubenswrapper[4777]: I1124 18:00:44.271898 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 18:00:44 crc kubenswrapper[4777]: I1124 18:00:44.272029 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" gracePeriod=600 Nov 24 18:00:44 crc kubenswrapper[4777]: E1124 18:00:44.407804 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:00:44 crc kubenswrapper[4777]: I1124 18:00:44.673476 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" exitCode=0 Nov 24 18:00:44 crc kubenswrapper[4777]: I1124 18:00:44.673520 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6"} Nov 24 18:00:44 crc kubenswrapper[4777]: I1124 18:00:44.673554 4777 scope.go:117] "RemoveContainer" containerID="4232c66077cf2691260b3a2ead0239df3749a3cf336096399a2eb7ed33863917" Nov 24 18:00:44 crc kubenswrapper[4777]: I1124 18:00:44.674292 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:00:44 crc kubenswrapper[4777]: E1124 18:00:44.674778 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:00:56 crc kubenswrapper[4777]: I1124 18:00:56.245392 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:00:56 crc kubenswrapper[4777]: E1124 18:00:56.246238 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.151999 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29400121-tzwwt"] Nov 24 18:01:00 crc kubenswrapper[4777]: E1124 18:01:00.153987 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b575ae4b-aac8-4a05-a096-a85cae8f1718" containerName="copy" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.154133 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="b575ae4b-aac8-4a05-a096-a85cae8f1718" containerName="copy" Nov 24 18:01:00 crc kubenswrapper[4777]: E1124 18:01:00.154234 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b575ae4b-aac8-4a05-a096-a85cae8f1718" containerName="gather" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.154299 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="b575ae4b-aac8-4a05-a096-a85cae8f1718" containerName="gather" Nov 24 18:01:00 crc kubenswrapper[4777]: E1124 18:01:00.154383 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cff5213-efb6-4d07-ae75-e9eb38583771" containerName="collect-profiles" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.154448 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cff5213-efb6-4d07-ae75-e9eb38583771" containerName="collect-profiles" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.154773 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="b575ae4b-aac8-4a05-a096-a85cae8f1718" containerName="gather" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.154879 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="b575ae4b-aac8-4a05-a096-a85cae8f1718" containerName="copy" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.154951 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cff5213-efb6-4d07-ae75-e9eb38583771" containerName="collect-profiles" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.156111 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.172832 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29400121-tzwwt"] Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.264451 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-fernet-keys\") pod \"keystone-cron-29400121-tzwwt\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.264521 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-config-data\") pod \"keystone-cron-29400121-tzwwt\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.264797 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-combined-ca-bundle\") pod \"keystone-cron-29400121-tzwwt\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.265007 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p27d\" (UniqueName: \"kubernetes.io/projected/33e57905-033a-4880-84e8-fe2d92272f5b-kube-api-access-7p27d\") pod \"keystone-cron-29400121-tzwwt\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.368254 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-config-data\") pod \"keystone-cron-29400121-tzwwt\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.368528 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-combined-ca-bundle\") pod \"keystone-cron-29400121-tzwwt\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.368581 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p27d\" (UniqueName: \"kubernetes.io/projected/33e57905-033a-4880-84e8-fe2d92272f5b-kube-api-access-7p27d\") pod \"keystone-cron-29400121-tzwwt\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.368628 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-fernet-keys\") pod \"keystone-cron-29400121-tzwwt\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.376228 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-config-data\") pod \"keystone-cron-29400121-tzwwt\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.376858 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-fernet-keys\") pod \"keystone-cron-29400121-tzwwt\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.386486 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-combined-ca-bundle\") pod \"keystone-cron-29400121-tzwwt\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.400852 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p27d\" (UniqueName: \"kubernetes.io/projected/33e57905-033a-4880-84e8-fe2d92272f5b-kube-api-access-7p27d\") pod \"keystone-cron-29400121-tzwwt\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.487949 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:00 crc kubenswrapper[4777]: I1124 18:01:00.960637 4777 scope.go:117] "RemoveContainer" containerID="ec4843ab252ebb79bcc5b433b19d8e6a94adb0eebb269d8a732c4edf6fbf5120" Nov 24 18:01:01 crc kubenswrapper[4777]: I1124 18:01:01.058153 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29400121-tzwwt"] Nov 24 18:01:01 crc kubenswrapper[4777]: I1124 18:01:01.864423 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400121-tzwwt" event={"ID":"33e57905-033a-4880-84e8-fe2d92272f5b","Type":"ContainerStarted","Data":"a7102e55a6a01ae6db873daca78b70f4800cbe497ec3f36015d1d1c27b09be0b"} Nov 24 18:01:01 crc kubenswrapper[4777]: I1124 18:01:01.864932 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400121-tzwwt" event={"ID":"33e57905-033a-4880-84e8-fe2d92272f5b","Type":"ContainerStarted","Data":"a8a0f0bf2861cbf4ae89c97e20c058e7d197cccc61a15467b1ec63b7ab39b1a0"} Nov 24 18:01:01 crc kubenswrapper[4777]: I1124 18:01:01.882782 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29400121-tzwwt" podStartSLOduration=1.882766872 podStartE2EDuration="1.882766872s" podCreationTimestamp="2025-11-24 18:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 18:01:01.882268978 +0000 UTC m=+3620.041464037" watchObservedRunningTime="2025-11-24 18:01:01.882766872 +0000 UTC m=+3620.041961921" Nov 24 18:01:04 crc kubenswrapper[4777]: I1124 18:01:04.892796 4777 generic.go:334] "Generic (PLEG): container finished" podID="33e57905-033a-4880-84e8-fe2d92272f5b" containerID="a7102e55a6a01ae6db873daca78b70f4800cbe497ec3f36015d1d1c27b09be0b" exitCode=0 Nov 24 18:01:04 crc kubenswrapper[4777]: I1124 18:01:04.892830 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400121-tzwwt" event={"ID":"33e57905-033a-4880-84e8-fe2d92272f5b","Type":"ContainerDied","Data":"a7102e55a6a01ae6db873daca78b70f4800cbe497ec3f36015d1d1c27b09be0b"} Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.561895 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.749953 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-config-data\") pod \"33e57905-033a-4880-84e8-fe2d92272f5b\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.750355 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p27d\" (UniqueName: \"kubernetes.io/projected/33e57905-033a-4880-84e8-fe2d92272f5b-kube-api-access-7p27d\") pod \"33e57905-033a-4880-84e8-fe2d92272f5b\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.750416 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-fernet-keys\") pod \"33e57905-033a-4880-84e8-fe2d92272f5b\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.750568 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-combined-ca-bundle\") pod \"33e57905-033a-4880-84e8-fe2d92272f5b\" (UID: \"33e57905-033a-4880-84e8-fe2d92272f5b\") " Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.756344 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "33e57905-033a-4880-84e8-fe2d92272f5b" (UID: "33e57905-033a-4880-84e8-fe2d92272f5b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.758016 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33e57905-033a-4880-84e8-fe2d92272f5b-kube-api-access-7p27d" (OuterVolumeSpecName: "kube-api-access-7p27d") pod "33e57905-033a-4880-84e8-fe2d92272f5b" (UID: "33e57905-033a-4880-84e8-fe2d92272f5b"). InnerVolumeSpecName "kube-api-access-7p27d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.789952 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33e57905-033a-4880-84e8-fe2d92272f5b" (UID: "33e57905-033a-4880-84e8-fe2d92272f5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.823574 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-config-data" (OuterVolumeSpecName: "config-data") pod "33e57905-033a-4880-84e8-fe2d92272f5b" (UID: "33e57905-033a-4880-84e8-fe2d92272f5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.853200 4777 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.853247 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p27d\" (UniqueName: \"kubernetes.io/projected/33e57905-033a-4880-84e8-fe2d92272f5b-kube-api-access-7p27d\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.853264 4777 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.853274 4777 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e57905-033a-4880-84e8-fe2d92272f5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.916753 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400121-tzwwt" event={"ID":"33e57905-033a-4880-84e8-fe2d92272f5b","Type":"ContainerDied","Data":"a8a0f0bf2861cbf4ae89c97e20c058e7d197cccc61a15467b1ec63b7ab39b1a0"} Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.916796 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8a0f0bf2861cbf4ae89c97e20c058e7d197cccc61a15467b1ec63b7ab39b1a0" Nov 24 18:01:06 crc kubenswrapper[4777]: I1124 18:01:06.916812 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400121-tzwwt" Nov 24 18:01:07 crc kubenswrapper[4777]: I1124 18:01:07.244839 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:01:07 crc kubenswrapper[4777]: E1124 18:01:07.245274 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:01:19 crc kubenswrapper[4777]: I1124 18:01:19.246565 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:01:19 crc kubenswrapper[4777]: E1124 18:01:19.247443 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:01:20 crc kubenswrapper[4777]: I1124 18:01:20.279077 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mznxc"] Nov 24 18:01:20 crc kubenswrapper[4777]: E1124 18:01:20.279692 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33e57905-033a-4880-84e8-fe2d92272f5b" containerName="keystone-cron" Nov 24 18:01:20 crc kubenswrapper[4777]: I1124 18:01:20.279707 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="33e57905-033a-4880-84e8-fe2d92272f5b" containerName="keystone-cron" Nov 24 18:01:20 crc kubenswrapper[4777]: I1124 18:01:20.280017 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="33e57905-033a-4880-84e8-fe2d92272f5b" containerName="keystone-cron" Nov 24 18:01:20 crc kubenswrapper[4777]: I1124 18:01:20.282174 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:20 crc kubenswrapper[4777]: I1124 18:01:20.290259 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mznxc"] Nov 24 18:01:20 crc kubenswrapper[4777]: I1124 18:01:20.421939 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-utilities\") pod \"redhat-marketplace-mznxc\" (UID: \"185cebfe-0b74-4528-9dcc-47daf0b6f7b8\") " pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:20 crc kubenswrapper[4777]: I1124 18:01:20.422325 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zd9p\" (UniqueName: \"kubernetes.io/projected/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-kube-api-access-8zd9p\") pod \"redhat-marketplace-mznxc\" (UID: \"185cebfe-0b74-4528-9dcc-47daf0b6f7b8\") " pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:20 crc kubenswrapper[4777]: I1124 18:01:20.422419 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-catalog-content\") pod \"redhat-marketplace-mznxc\" (UID: \"185cebfe-0b74-4528-9dcc-47daf0b6f7b8\") " pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:20 crc kubenswrapper[4777]: I1124 18:01:20.524743 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-catalog-content\") pod \"redhat-marketplace-mznxc\" (UID: \"185cebfe-0b74-4528-9dcc-47daf0b6f7b8\") " pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:20 crc kubenswrapper[4777]: I1124 18:01:20.524914 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-utilities\") pod \"redhat-marketplace-mznxc\" (UID: \"185cebfe-0b74-4528-9dcc-47daf0b6f7b8\") " pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:20 crc kubenswrapper[4777]: I1124 18:01:20.524950 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zd9p\" (UniqueName: \"kubernetes.io/projected/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-kube-api-access-8zd9p\") pod \"redhat-marketplace-mznxc\" (UID: \"185cebfe-0b74-4528-9dcc-47daf0b6f7b8\") " pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:20 crc kubenswrapper[4777]: I1124 18:01:20.525343 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-catalog-content\") pod \"redhat-marketplace-mznxc\" (UID: \"185cebfe-0b74-4528-9dcc-47daf0b6f7b8\") " pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:20 crc kubenswrapper[4777]: I1124 18:01:20.525362 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-utilities\") pod \"redhat-marketplace-mznxc\" (UID: \"185cebfe-0b74-4528-9dcc-47daf0b6f7b8\") " pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:20 crc kubenswrapper[4777]: I1124 18:01:20.551797 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zd9p\" (UniqueName: \"kubernetes.io/projected/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-kube-api-access-8zd9p\") pod \"redhat-marketplace-mznxc\" (UID: \"185cebfe-0b74-4528-9dcc-47daf0b6f7b8\") " pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:20 crc kubenswrapper[4777]: I1124 18:01:20.613613 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:21 crc kubenswrapper[4777]: I1124 18:01:21.242654 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mznxc"] Nov 24 18:01:22 crc kubenswrapper[4777]: I1124 18:01:22.078061 4777 generic.go:334] "Generic (PLEG): container finished" podID="185cebfe-0b74-4528-9dcc-47daf0b6f7b8" containerID="d7a23c51b2dcb2525af493d5e657a68d7df24b435892547935c2e1323b7dde00" exitCode=0 Nov 24 18:01:22 crc kubenswrapper[4777]: I1124 18:01:22.078320 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mznxc" event={"ID":"185cebfe-0b74-4528-9dcc-47daf0b6f7b8","Type":"ContainerDied","Data":"d7a23c51b2dcb2525af493d5e657a68d7df24b435892547935c2e1323b7dde00"} Nov 24 18:01:22 crc kubenswrapper[4777]: I1124 18:01:22.078425 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mznxc" event={"ID":"185cebfe-0b74-4528-9dcc-47daf0b6f7b8","Type":"ContainerStarted","Data":"140efa140da41edd039543236d8496512fc20eb15332787ed1ffcecdb56b7e95"} Nov 24 18:01:22 crc kubenswrapper[4777]: I1124 18:01:22.081001 4777 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 18:01:24 crc kubenswrapper[4777]: I1124 18:01:24.103103 4777 generic.go:334] "Generic (PLEG): container finished" podID="185cebfe-0b74-4528-9dcc-47daf0b6f7b8" containerID="3f257f2773a8c5c99427bff0f816721233a705a74e7e0a3588b55d50d7369238" exitCode=0 Nov 24 18:01:24 crc kubenswrapper[4777]: I1124 18:01:24.103202 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mznxc" event={"ID":"185cebfe-0b74-4528-9dcc-47daf0b6f7b8","Type":"ContainerDied","Data":"3f257f2773a8c5c99427bff0f816721233a705a74e7e0a3588b55d50d7369238"} Nov 24 18:01:25 crc kubenswrapper[4777]: I1124 18:01:25.115582 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mznxc" event={"ID":"185cebfe-0b74-4528-9dcc-47daf0b6f7b8","Type":"ContainerStarted","Data":"1439e6ee8714979391ed29e894a283af90f11960abd12cc4fafedbc9230364de"} Nov 24 18:01:25 crc kubenswrapper[4777]: I1124 18:01:25.141808 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mznxc" podStartSLOduration=2.458732952 podStartE2EDuration="5.141771133s" podCreationTimestamp="2025-11-24 18:01:20 +0000 UTC" firstStartedPulling="2025-11-24 18:01:22.080714571 +0000 UTC m=+3640.239909630" lastFinishedPulling="2025-11-24 18:01:24.763752762 +0000 UTC m=+3642.922947811" observedRunningTime="2025-11-24 18:01:25.13358977 +0000 UTC m=+3643.292784829" watchObservedRunningTime="2025-11-24 18:01:25.141771133 +0000 UTC m=+3643.300966182" Nov 24 18:01:30 crc kubenswrapper[4777]: I1124 18:01:30.614649 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:30 crc kubenswrapper[4777]: I1124 18:01:30.615113 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:30 crc kubenswrapper[4777]: I1124 18:01:30.701482 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:31 crc kubenswrapper[4777]: I1124 18:01:31.208663 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:31 crc kubenswrapper[4777]: I1124 18:01:31.263580 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:01:31 crc kubenswrapper[4777]: E1124 18:01:31.263853 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:01:31 crc kubenswrapper[4777]: I1124 18:01:31.272340 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mznxc"] Nov 24 18:01:33 crc kubenswrapper[4777]: I1124 18:01:33.188144 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mznxc" podUID="185cebfe-0b74-4528-9dcc-47daf0b6f7b8" containerName="registry-server" containerID="cri-o://1439e6ee8714979391ed29e894a283af90f11960abd12cc4fafedbc9230364de" gracePeriod=2 Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.040449 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.108858 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-catalog-content\") pod \"185cebfe-0b74-4528-9dcc-47daf0b6f7b8\" (UID: \"185cebfe-0b74-4528-9dcc-47daf0b6f7b8\") " Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.109005 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-utilities\") pod \"185cebfe-0b74-4528-9dcc-47daf0b6f7b8\" (UID: \"185cebfe-0b74-4528-9dcc-47daf0b6f7b8\") " Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.109065 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zd9p\" (UniqueName: \"kubernetes.io/projected/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-kube-api-access-8zd9p\") pod \"185cebfe-0b74-4528-9dcc-47daf0b6f7b8\" (UID: \"185cebfe-0b74-4528-9dcc-47daf0b6f7b8\") " Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.110777 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-utilities" (OuterVolumeSpecName: "utilities") pod "185cebfe-0b74-4528-9dcc-47daf0b6f7b8" (UID: "185cebfe-0b74-4528-9dcc-47daf0b6f7b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.121712 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-kube-api-access-8zd9p" (OuterVolumeSpecName: "kube-api-access-8zd9p") pod "185cebfe-0b74-4528-9dcc-47daf0b6f7b8" (UID: "185cebfe-0b74-4528-9dcc-47daf0b6f7b8"). InnerVolumeSpecName "kube-api-access-8zd9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.131832 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "185cebfe-0b74-4528-9dcc-47daf0b6f7b8" (UID: "185cebfe-0b74-4528-9dcc-47daf0b6f7b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.204694 4777 generic.go:334] "Generic (PLEG): container finished" podID="185cebfe-0b74-4528-9dcc-47daf0b6f7b8" containerID="1439e6ee8714979391ed29e894a283af90f11960abd12cc4fafedbc9230364de" exitCode=0 Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.205078 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mznxc" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.205085 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mznxc" event={"ID":"185cebfe-0b74-4528-9dcc-47daf0b6f7b8","Type":"ContainerDied","Data":"1439e6ee8714979391ed29e894a283af90f11960abd12cc4fafedbc9230364de"} Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.205140 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mznxc" event={"ID":"185cebfe-0b74-4528-9dcc-47daf0b6f7b8","Type":"ContainerDied","Data":"140efa140da41edd039543236d8496512fc20eb15332787ed1ffcecdb56b7e95"} Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.205160 4777 scope.go:117] "RemoveContainer" containerID="1439e6ee8714979391ed29e894a283af90f11960abd12cc4fafedbc9230364de" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.214875 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.214897 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.214908 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zd9p\" (UniqueName: \"kubernetes.io/projected/185cebfe-0b74-4528-9dcc-47daf0b6f7b8-kube-api-access-8zd9p\") on node \"crc\" DevicePath \"\"" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.229305 4777 scope.go:117] "RemoveContainer" containerID="3f257f2773a8c5c99427bff0f816721233a705a74e7e0a3588b55d50d7369238" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.249326 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mznxc"] Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.257916 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mznxc"] Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.272407 4777 scope.go:117] "RemoveContainer" containerID="d7a23c51b2dcb2525af493d5e657a68d7df24b435892547935c2e1323b7dde00" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.305404 4777 scope.go:117] "RemoveContainer" containerID="1439e6ee8714979391ed29e894a283af90f11960abd12cc4fafedbc9230364de" Nov 24 18:01:34 crc kubenswrapper[4777]: E1124 18:01:34.305950 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1439e6ee8714979391ed29e894a283af90f11960abd12cc4fafedbc9230364de\": container with ID starting with 1439e6ee8714979391ed29e894a283af90f11960abd12cc4fafedbc9230364de not found: ID does not exist" containerID="1439e6ee8714979391ed29e894a283af90f11960abd12cc4fafedbc9230364de" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.306014 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1439e6ee8714979391ed29e894a283af90f11960abd12cc4fafedbc9230364de"} err="failed to get container status \"1439e6ee8714979391ed29e894a283af90f11960abd12cc4fafedbc9230364de\": rpc error: code = NotFound desc = could not find container \"1439e6ee8714979391ed29e894a283af90f11960abd12cc4fafedbc9230364de\": container with ID starting with 1439e6ee8714979391ed29e894a283af90f11960abd12cc4fafedbc9230364de not found: ID does not exist" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.306049 4777 scope.go:117] "RemoveContainer" containerID="3f257f2773a8c5c99427bff0f816721233a705a74e7e0a3588b55d50d7369238" Nov 24 18:01:34 crc kubenswrapper[4777]: E1124 18:01:34.306451 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f257f2773a8c5c99427bff0f816721233a705a74e7e0a3588b55d50d7369238\": container with ID starting with 3f257f2773a8c5c99427bff0f816721233a705a74e7e0a3588b55d50d7369238 not found: ID does not exist" containerID="3f257f2773a8c5c99427bff0f816721233a705a74e7e0a3588b55d50d7369238" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.306490 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f257f2773a8c5c99427bff0f816721233a705a74e7e0a3588b55d50d7369238"} err="failed to get container status \"3f257f2773a8c5c99427bff0f816721233a705a74e7e0a3588b55d50d7369238\": rpc error: code = NotFound desc = could not find container \"3f257f2773a8c5c99427bff0f816721233a705a74e7e0a3588b55d50d7369238\": container with ID starting with 3f257f2773a8c5c99427bff0f816721233a705a74e7e0a3588b55d50d7369238 not found: ID does not exist" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.306517 4777 scope.go:117] "RemoveContainer" containerID="d7a23c51b2dcb2525af493d5e657a68d7df24b435892547935c2e1323b7dde00" Nov 24 18:01:34 crc kubenswrapper[4777]: E1124 18:01:34.306906 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7a23c51b2dcb2525af493d5e657a68d7df24b435892547935c2e1323b7dde00\": container with ID starting with d7a23c51b2dcb2525af493d5e657a68d7df24b435892547935c2e1323b7dde00 not found: ID does not exist" containerID="d7a23c51b2dcb2525af493d5e657a68d7df24b435892547935c2e1323b7dde00" Nov 24 18:01:34 crc kubenswrapper[4777]: I1124 18:01:34.306924 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7a23c51b2dcb2525af493d5e657a68d7df24b435892547935c2e1323b7dde00"} err="failed to get container status \"d7a23c51b2dcb2525af493d5e657a68d7df24b435892547935c2e1323b7dde00\": rpc error: code = NotFound desc = could not find container \"d7a23c51b2dcb2525af493d5e657a68d7df24b435892547935c2e1323b7dde00\": container with ID starting with d7a23c51b2dcb2525af493d5e657a68d7df24b435892547935c2e1323b7dde00 not found: ID does not exist" Nov 24 18:01:35 crc kubenswrapper[4777]: I1124 18:01:35.257299 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="185cebfe-0b74-4528-9dcc-47daf0b6f7b8" path="/var/lib/kubelet/pods/185cebfe-0b74-4528-9dcc-47daf0b6f7b8/volumes" Nov 24 18:01:44 crc kubenswrapper[4777]: I1124 18:01:44.246125 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:01:44 crc kubenswrapper[4777]: E1124 18:01:44.246948 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:01:59 crc kubenswrapper[4777]: I1124 18:01:59.245119 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:01:59 crc kubenswrapper[4777]: E1124 18:01:59.245921 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:02:01 crc kubenswrapper[4777]: I1124 18:02:01.035773 4777 scope.go:117] "RemoveContainer" containerID="e0203afc708a619716fcce5fafa3d9fe7cb926c114f86b2f7e40125c3b3e18b0" Nov 24 18:02:13 crc kubenswrapper[4777]: I1124 18:02:13.254810 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:02:13 crc kubenswrapper[4777]: E1124 18:02:13.255611 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:02:27 crc kubenswrapper[4777]: I1124 18:02:27.245810 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:02:27 crc kubenswrapper[4777]: E1124 18:02:27.246901 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:02:41 crc kubenswrapper[4777]: I1124 18:02:41.245954 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:02:41 crc kubenswrapper[4777]: E1124 18:02:41.246921 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:02:53 crc kubenswrapper[4777]: I1124 18:02:53.257204 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:02:53 crc kubenswrapper[4777]: E1124 18:02:53.258053 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:03:05 crc kubenswrapper[4777]: I1124 18:03:05.245617 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:03:05 crc kubenswrapper[4777]: E1124 18:03:05.246413 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:03:19 crc kubenswrapper[4777]: I1124 18:03:19.248670 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:03:19 crc kubenswrapper[4777]: E1124 18:03:19.249420 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:03:34 crc kubenswrapper[4777]: I1124 18:03:34.245906 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:03:34 crc kubenswrapper[4777]: E1124 18:03:34.247004 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:03:41 crc kubenswrapper[4777]: I1124 18:03:41.732203 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-w8n5g/must-gather-tvr92"] Nov 24 18:03:41 crc kubenswrapper[4777]: E1124 18:03:41.733135 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185cebfe-0b74-4528-9dcc-47daf0b6f7b8" containerName="registry-server" Nov 24 18:03:41 crc kubenswrapper[4777]: I1124 18:03:41.733148 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="185cebfe-0b74-4528-9dcc-47daf0b6f7b8" containerName="registry-server" Nov 24 18:03:41 crc kubenswrapper[4777]: E1124 18:03:41.733164 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185cebfe-0b74-4528-9dcc-47daf0b6f7b8" containerName="extract-content" Nov 24 18:03:41 crc kubenswrapper[4777]: I1124 18:03:41.733170 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="185cebfe-0b74-4528-9dcc-47daf0b6f7b8" containerName="extract-content" Nov 24 18:03:41 crc kubenswrapper[4777]: E1124 18:03:41.733190 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185cebfe-0b74-4528-9dcc-47daf0b6f7b8" containerName="extract-utilities" Nov 24 18:03:41 crc kubenswrapper[4777]: I1124 18:03:41.733197 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="185cebfe-0b74-4528-9dcc-47daf0b6f7b8" containerName="extract-utilities" Nov 24 18:03:41 crc kubenswrapper[4777]: I1124 18:03:41.733393 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="185cebfe-0b74-4528-9dcc-47daf0b6f7b8" containerName="registry-server" Nov 24 18:03:41 crc kubenswrapper[4777]: I1124 18:03:41.743292 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/must-gather-tvr92" Nov 24 18:03:41 crc kubenswrapper[4777]: I1124 18:03:41.745059 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-w8n5g"/"openshift-service-ca.crt" Nov 24 18:03:41 crc kubenswrapper[4777]: I1124 18:03:41.753640 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-w8n5g/must-gather-tvr92"] Nov 24 18:03:41 crc kubenswrapper[4777]: I1124 18:03:41.754327 4777 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-w8n5g"/"kube-root-ca.crt" Nov 24 18:03:41 crc kubenswrapper[4777]: I1124 18:03:41.856204 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llvdc\" (UniqueName: \"kubernetes.io/projected/a0ef86d3-63fe-421c-b274-eee544924006-kube-api-access-llvdc\") pod \"must-gather-tvr92\" (UID: \"a0ef86d3-63fe-421c-b274-eee544924006\") " pod="openshift-must-gather-w8n5g/must-gather-tvr92" Nov 24 18:03:41 crc kubenswrapper[4777]: I1124 18:03:41.857615 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a0ef86d3-63fe-421c-b274-eee544924006-must-gather-output\") pod \"must-gather-tvr92\" (UID: \"a0ef86d3-63fe-421c-b274-eee544924006\") " pod="openshift-must-gather-w8n5g/must-gather-tvr92" Nov 24 18:03:41 crc kubenswrapper[4777]: I1124 18:03:41.959017 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a0ef86d3-63fe-421c-b274-eee544924006-must-gather-output\") pod \"must-gather-tvr92\" (UID: \"a0ef86d3-63fe-421c-b274-eee544924006\") " pod="openshift-must-gather-w8n5g/must-gather-tvr92" Nov 24 18:03:41 crc kubenswrapper[4777]: I1124 18:03:41.959121 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llvdc\" (UniqueName: \"kubernetes.io/projected/a0ef86d3-63fe-421c-b274-eee544924006-kube-api-access-llvdc\") pod \"must-gather-tvr92\" (UID: \"a0ef86d3-63fe-421c-b274-eee544924006\") " pod="openshift-must-gather-w8n5g/must-gather-tvr92" Nov 24 18:03:41 crc kubenswrapper[4777]: I1124 18:03:41.959580 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a0ef86d3-63fe-421c-b274-eee544924006-must-gather-output\") pod \"must-gather-tvr92\" (UID: \"a0ef86d3-63fe-421c-b274-eee544924006\") " pod="openshift-must-gather-w8n5g/must-gather-tvr92" Nov 24 18:03:41 crc kubenswrapper[4777]: I1124 18:03:41.981081 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llvdc\" (UniqueName: \"kubernetes.io/projected/a0ef86d3-63fe-421c-b274-eee544924006-kube-api-access-llvdc\") pod \"must-gather-tvr92\" (UID: \"a0ef86d3-63fe-421c-b274-eee544924006\") " pod="openshift-must-gather-w8n5g/must-gather-tvr92" Nov 24 18:03:42 crc kubenswrapper[4777]: I1124 18:03:42.076672 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/must-gather-tvr92" Nov 24 18:03:42 crc kubenswrapper[4777]: I1124 18:03:42.616134 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-w8n5g/must-gather-tvr92"] Nov 24 18:03:43 crc kubenswrapper[4777]: I1124 18:03:43.628305 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w8n5g/must-gather-tvr92" event={"ID":"a0ef86d3-63fe-421c-b274-eee544924006","Type":"ContainerStarted","Data":"7232f211d58149ba1bf05b71c2cff6ccd64cc2c20ca455d6d9afb5dc50827621"} Nov 24 18:03:43 crc kubenswrapper[4777]: I1124 18:03:43.628888 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w8n5g/must-gather-tvr92" event={"ID":"a0ef86d3-63fe-421c-b274-eee544924006","Type":"ContainerStarted","Data":"a5a2aca4ed00f1883aa49419845e5ddcf3ef7432e5e6ad690124e5a7fecb7b93"} Nov 24 18:03:43 crc kubenswrapper[4777]: I1124 18:03:43.628905 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w8n5g/must-gather-tvr92" event={"ID":"a0ef86d3-63fe-421c-b274-eee544924006","Type":"ContainerStarted","Data":"ef197b5bfa411302e1f0e819205377703552af3b919d6172b6aa01c883e657bd"} Nov 24 18:03:43 crc kubenswrapper[4777]: I1124 18:03:43.673303 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-w8n5g/must-gather-tvr92" podStartSLOduration=2.6732770820000002 podStartE2EDuration="2.673277082s" podCreationTimestamp="2025-11-24 18:03:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 18:03:43.640796756 +0000 UTC m=+3781.799991805" watchObservedRunningTime="2025-11-24 18:03:43.673277082 +0000 UTC m=+3781.832472151" Nov 24 18:03:46 crc kubenswrapper[4777]: I1124 18:03:46.245331 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:03:46 crc kubenswrapper[4777]: E1124 18:03:46.246045 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:03:46 crc kubenswrapper[4777]: I1124 18:03:46.690506 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-w8n5g/crc-debug-h7nk4"] Nov 24 18:03:46 crc kubenswrapper[4777]: I1124 18:03:46.692196 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/crc-debug-h7nk4" Nov 24 18:03:46 crc kubenswrapper[4777]: I1124 18:03:46.695277 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-w8n5g"/"default-dockercfg-vt6nh" Nov 24 18:03:46 crc kubenswrapper[4777]: I1124 18:03:46.766263 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z77xp\" (UniqueName: \"kubernetes.io/projected/34cbc796-d4f6-44ee-9d9a-30465d08bdc6-kube-api-access-z77xp\") pod \"crc-debug-h7nk4\" (UID: \"34cbc796-d4f6-44ee-9d9a-30465d08bdc6\") " pod="openshift-must-gather-w8n5g/crc-debug-h7nk4" Nov 24 18:03:46 crc kubenswrapper[4777]: I1124 18:03:46.766744 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34cbc796-d4f6-44ee-9d9a-30465d08bdc6-host\") pod \"crc-debug-h7nk4\" (UID: \"34cbc796-d4f6-44ee-9d9a-30465d08bdc6\") " pod="openshift-must-gather-w8n5g/crc-debug-h7nk4" Nov 24 18:03:46 crc kubenswrapper[4777]: I1124 18:03:46.869558 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34cbc796-d4f6-44ee-9d9a-30465d08bdc6-host\") pod \"crc-debug-h7nk4\" (UID: \"34cbc796-d4f6-44ee-9d9a-30465d08bdc6\") " pod="openshift-must-gather-w8n5g/crc-debug-h7nk4" Nov 24 18:03:46 crc kubenswrapper[4777]: I1124 18:03:46.870217 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z77xp\" (UniqueName: \"kubernetes.io/projected/34cbc796-d4f6-44ee-9d9a-30465d08bdc6-kube-api-access-z77xp\") pod \"crc-debug-h7nk4\" (UID: \"34cbc796-d4f6-44ee-9d9a-30465d08bdc6\") " pod="openshift-must-gather-w8n5g/crc-debug-h7nk4" Nov 24 18:03:46 crc kubenswrapper[4777]: I1124 18:03:46.869700 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34cbc796-d4f6-44ee-9d9a-30465d08bdc6-host\") pod \"crc-debug-h7nk4\" (UID: \"34cbc796-d4f6-44ee-9d9a-30465d08bdc6\") " pod="openshift-must-gather-w8n5g/crc-debug-h7nk4" Nov 24 18:03:46 crc kubenswrapper[4777]: I1124 18:03:46.920311 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z77xp\" (UniqueName: \"kubernetes.io/projected/34cbc796-d4f6-44ee-9d9a-30465d08bdc6-kube-api-access-z77xp\") pod \"crc-debug-h7nk4\" (UID: \"34cbc796-d4f6-44ee-9d9a-30465d08bdc6\") " pod="openshift-must-gather-w8n5g/crc-debug-h7nk4" Nov 24 18:03:47 crc kubenswrapper[4777]: I1124 18:03:47.016772 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/crc-debug-h7nk4" Nov 24 18:03:47 crc kubenswrapper[4777]: W1124 18:03:47.059754 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34cbc796_d4f6_44ee_9d9a_30465d08bdc6.slice/crio-f0a324d7e7c2e2e637ab2f68c3b7221536fad166565ddafbb81438081396eb82 WatchSource:0}: Error finding container f0a324d7e7c2e2e637ab2f68c3b7221536fad166565ddafbb81438081396eb82: Status 404 returned error can't find the container with id f0a324d7e7c2e2e637ab2f68c3b7221536fad166565ddafbb81438081396eb82 Nov 24 18:03:47 crc kubenswrapper[4777]: I1124 18:03:47.673135 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w8n5g/crc-debug-h7nk4" event={"ID":"34cbc796-d4f6-44ee-9d9a-30465d08bdc6","Type":"ContainerStarted","Data":"3604715c19080e139ef8fdae4ce5ed671a6ed6d3a8c690fc23fc2f851556f967"} Nov 24 18:03:47 crc kubenswrapper[4777]: I1124 18:03:47.673613 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w8n5g/crc-debug-h7nk4" event={"ID":"34cbc796-d4f6-44ee-9d9a-30465d08bdc6","Type":"ContainerStarted","Data":"f0a324d7e7c2e2e637ab2f68c3b7221536fad166565ddafbb81438081396eb82"} Nov 24 18:03:47 crc kubenswrapper[4777]: I1124 18:03:47.694728 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-w8n5g/crc-debug-h7nk4" podStartSLOduration=1.694710274 podStartE2EDuration="1.694710274s" podCreationTimestamp="2025-11-24 18:03:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 18:03:47.68650104 +0000 UTC m=+3785.845696089" watchObservedRunningTime="2025-11-24 18:03:47.694710274 +0000 UTC m=+3785.853905323" Nov 24 18:04:01 crc kubenswrapper[4777]: I1124 18:04:01.245081 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:04:01 crc kubenswrapper[4777]: E1124 18:04:01.246063 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:04:15 crc kubenswrapper[4777]: I1124 18:04:15.245291 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:04:15 crc kubenswrapper[4777]: E1124 18:04:15.246142 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:04:29 crc kubenswrapper[4777]: I1124 18:04:29.244797 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:04:29 crc kubenswrapper[4777]: E1124 18:04:29.245566 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:04:35 crc kubenswrapper[4777]: I1124 18:04:35.124561 4777 generic.go:334] "Generic (PLEG): container finished" podID="34cbc796-d4f6-44ee-9d9a-30465d08bdc6" containerID="3604715c19080e139ef8fdae4ce5ed671a6ed6d3a8c690fc23fc2f851556f967" exitCode=0 Nov 24 18:04:35 crc kubenswrapper[4777]: I1124 18:04:35.124939 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w8n5g/crc-debug-h7nk4" event={"ID":"34cbc796-d4f6-44ee-9d9a-30465d08bdc6","Type":"ContainerDied","Data":"3604715c19080e139ef8fdae4ce5ed671a6ed6d3a8c690fc23fc2f851556f967"} Nov 24 18:04:36 crc kubenswrapper[4777]: I1124 18:04:36.258694 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/crc-debug-h7nk4" Nov 24 18:04:36 crc kubenswrapper[4777]: I1124 18:04:36.300597 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-w8n5g/crc-debug-h7nk4"] Nov 24 18:04:36 crc kubenswrapper[4777]: I1124 18:04:36.309021 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-w8n5g/crc-debug-h7nk4"] Nov 24 18:04:36 crc kubenswrapper[4777]: I1124 18:04:36.401940 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34cbc796-d4f6-44ee-9d9a-30465d08bdc6-host\") pod \"34cbc796-d4f6-44ee-9d9a-30465d08bdc6\" (UID: \"34cbc796-d4f6-44ee-9d9a-30465d08bdc6\") " Nov 24 18:04:36 crc kubenswrapper[4777]: I1124 18:04:36.402071 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34cbc796-d4f6-44ee-9d9a-30465d08bdc6-host" (OuterVolumeSpecName: "host") pod "34cbc796-d4f6-44ee-9d9a-30465d08bdc6" (UID: "34cbc796-d4f6-44ee-9d9a-30465d08bdc6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 18:04:36 crc kubenswrapper[4777]: I1124 18:04:36.402185 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z77xp\" (UniqueName: \"kubernetes.io/projected/34cbc796-d4f6-44ee-9d9a-30465d08bdc6-kube-api-access-z77xp\") pod \"34cbc796-d4f6-44ee-9d9a-30465d08bdc6\" (UID: \"34cbc796-d4f6-44ee-9d9a-30465d08bdc6\") " Nov 24 18:04:36 crc kubenswrapper[4777]: I1124 18:04:36.403120 4777 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34cbc796-d4f6-44ee-9d9a-30465d08bdc6-host\") on node \"crc\" DevicePath \"\"" Nov 24 18:04:36 crc kubenswrapper[4777]: I1124 18:04:36.410257 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34cbc796-d4f6-44ee-9d9a-30465d08bdc6-kube-api-access-z77xp" (OuterVolumeSpecName: "kube-api-access-z77xp") pod "34cbc796-d4f6-44ee-9d9a-30465d08bdc6" (UID: "34cbc796-d4f6-44ee-9d9a-30465d08bdc6"). InnerVolumeSpecName "kube-api-access-z77xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:04:36 crc kubenswrapper[4777]: I1124 18:04:36.504829 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z77xp\" (UniqueName: \"kubernetes.io/projected/34cbc796-d4f6-44ee-9d9a-30465d08bdc6-kube-api-access-z77xp\") on node \"crc\" DevicePath \"\"" Nov 24 18:04:37 crc kubenswrapper[4777]: I1124 18:04:37.146144 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0a324d7e7c2e2e637ab2f68c3b7221536fad166565ddafbb81438081396eb82" Nov 24 18:04:37 crc kubenswrapper[4777]: I1124 18:04:37.146198 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/crc-debug-h7nk4" Nov 24 18:04:37 crc kubenswrapper[4777]: I1124 18:04:37.255907 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34cbc796-d4f6-44ee-9d9a-30465d08bdc6" path="/var/lib/kubelet/pods/34cbc796-d4f6-44ee-9d9a-30465d08bdc6/volumes" Nov 24 18:04:37 crc kubenswrapper[4777]: I1124 18:04:37.491188 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-w8n5g/crc-debug-jjztl"] Nov 24 18:04:37 crc kubenswrapper[4777]: E1124 18:04:37.491735 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34cbc796-d4f6-44ee-9d9a-30465d08bdc6" containerName="container-00" Nov 24 18:04:37 crc kubenswrapper[4777]: I1124 18:04:37.491757 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="34cbc796-d4f6-44ee-9d9a-30465d08bdc6" containerName="container-00" Nov 24 18:04:37 crc kubenswrapper[4777]: I1124 18:04:37.492137 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="34cbc796-d4f6-44ee-9d9a-30465d08bdc6" containerName="container-00" Nov 24 18:04:37 crc kubenswrapper[4777]: I1124 18:04:37.493212 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/crc-debug-jjztl" Nov 24 18:04:37 crc kubenswrapper[4777]: I1124 18:04:37.495460 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-w8n5g"/"default-dockercfg-vt6nh" Nov 24 18:04:37 crc kubenswrapper[4777]: I1124 18:04:37.524884 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca02a1d6-3048-418e-9bbd-a872ce429218-host\") pod \"crc-debug-jjztl\" (UID: \"ca02a1d6-3048-418e-9bbd-a872ce429218\") " pod="openshift-must-gather-w8n5g/crc-debug-jjztl" Nov 24 18:04:37 crc kubenswrapper[4777]: I1124 18:04:37.525063 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hvxz\" (UniqueName: \"kubernetes.io/projected/ca02a1d6-3048-418e-9bbd-a872ce429218-kube-api-access-2hvxz\") pod \"crc-debug-jjztl\" (UID: \"ca02a1d6-3048-418e-9bbd-a872ce429218\") " pod="openshift-must-gather-w8n5g/crc-debug-jjztl" Nov 24 18:04:37 crc kubenswrapper[4777]: I1124 18:04:37.626743 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca02a1d6-3048-418e-9bbd-a872ce429218-host\") pod \"crc-debug-jjztl\" (UID: \"ca02a1d6-3048-418e-9bbd-a872ce429218\") " pod="openshift-must-gather-w8n5g/crc-debug-jjztl" Nov 24 18:04:37 crc kubenswrapper[4777]: I1124 18:04:37.626880 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hvxz\" (UniqueName: \"kubernetes.io/projected/ca02a1d6-3048-418e-9bbd-a872ce429218-kube-api-access-2hvxz\") pod \"crc-debug-jjztl\" (UID: \"ca02a1d6-3048-418e-9bbd-a872ce429218\") " pod="openshift-must-gather-w8n5g/crc-debug-jjztl" Nov 24 18:04:37 crc kubenswrapper[4777]: I1124 18:04:37.627356 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca02a1d6-3048-418e-9bbd-a872ce429218-host\") pod \"crc-debug-jjztl\" (UID: \"ca02a1d6-3048-418e-9bbd-a872ce429218\") " pod="openshift-must-gather-w8n5g/crc-debug-jjztl" Nov 24 18:04:37 crc kubenswrapper[4777]: I1124 18:04:37.645541 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hvxz\" (UniqueName: \"kubernetes.io/projected/ca02a1d6-3048-418e-9bbd-a872ce429218-kube-api-access-2hvxz\") pod \"crc-debug-jjztl\" (UID: \"ca02a1d6-3048-418e-9bbd-a872ce429218\") " pod="openshift-must-gather-w8n5g/crc-debug-jjztl" Nov 24 18:04:37 crc kubenswrapper[4777]: I1124 18:04:37.814003 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/crc-debug-jjztl" Nov 24 18:04:37 crc kubenswrapper[4777]: W1124 18:04:37.869388 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca02a1d6_3048_418e_9bbd_a872ce429218.slice/crio-ac409f93dae9c91835b20c3eeeb609fa8cb108b81528e203efb88537d29ea812 WatchSource:0}: Error finding container ac409f93dae9c91835b20c3eeeb609fa8cb108b81528e203efb88537d29ea812: Status 404 returned error can't find the container with id ac409f93dae9c91835b20c3eeeb609fa8cb108b81528e203efb88537d29ea812 Nov 24 18:04:38 crc kubenswrapper[4777]: I1124 18:04:38.160105 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w8n5g/crc-debug-jjztl" event={"ID":"ca02a1d6-3048-418e-9bbd-a872ce429218","Type":"ContainerStarted","Data":"c74f2257af9b0395bece23b5a83fb192cb3088590e1cee4c812edb1b9c71d517"} Nov 24 18:04:38 crc kubenswrapper[4777]: I1124 18:04:38.160445 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w8n5g/crc-debug-jjztl" event={"ID":"ca02a1d6-3048-418e-9bbd-a872ce429218","Type":"ContainerStarted","Data":"ac409f93dae9c91835b20c3eeeb609fa8cb108b81528e203efb88537d29ea812"} Nov 24 18:04:38 crc kubenswrapper[4777]: I1124 18:04:38.175223 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-w8n5g/crc-debug-jjztl" podStartSLOduration=1.175208034 podStartE2EDuration="1.175208034s" podCreationTimestamp="2025-11-24 18:04:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 18:04:38.170565332 +0000 UTC m=+3836.329760381" watchObservedRunningTime="2025-11-24 18:04:38.175208034 +0000 UTC m=+3836.334403083" Nov 24 18:04:39 crc kubenswrapper[4777]: I1124 18:04:39.174993 4777 generic.go:334] "Generic (PLEG): container finished" podID="ca02a1d6-3048-418e-9bbd-a872ce429218" containerID="c74f2257af9b0395bece23b5a83fb192cb3088590e1cee4c812edb1b9c71d517" exitCode=0 Nov 24 18:04:39 crc kubenswrapper[4777]: I1124 18:04:39.175334 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w8n5g/crc-debug-jjztl" event={"ID":"ca02a1d6-3048-418e-9bbd-a872ce429218","Type":"ContainerDied","Data":"c74f2257af9b0395bece23b5a83fb192cb3088590e1cee4c812edb1b9c71d517"} Nov 24 18:04:40 crc kubenswrapper[4777]: I1124 18:04:40.245381 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:04:40 crc kubenswrapper[4777]: E1124 18:04:40.245760 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:04:40 crc kubenswrapper[4777]: I1124 18:04:40.356814 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/crc-debug-jjztl" Nov 24 18:04:40 crc kubenswrapper[4777]: I1124 18:04:40.390692 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-w8n5g/crc-debug-jjztl"] Nov 24 18:04:40 crc kubenswrapper[4777]: I1124 18:04:40.403131 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-w8n5g/crc-debug-jjztl"] Nov 24 18:04:40 crc kubenswrapper[4777]: I1124 18:04:40.483106 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hvxz\" (UniqueName: \"kubernetes.io/projected/ca02a1d6-3048-418e-9bbd-a872ce429218-kube-api-access-2hvxz\") pod \"ca02a1d6-3048-418e-9bbd-a872ce429218\" (UID: \"ca02a1d6-3048-418e-9bbd-a872ce429218\") " Nov 24 18:04:40 crc kubenswrapper[4777]: I1124 18:04:40.483225 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca02a1d6-3048-418e-9bbd-a872ce429218-host\") pod \"ca02a1d6-3048-418e-9bbd-a872ce429218\" (UID: \"ca02a1d6-3048-418e-9bbd-a872ce429218\") " Nov 24 18:04:40 crc kubenswrapper[4777]: I1124 18:04:40.483322 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca02a1d6-3048-418e-9bbd-a872ce429218-host" (OuterVolumeSpecName: "host") pod "ca02a1d6-3048-418e-9bbd-a872ce429218" (UID: "ca02a1d6-3048-418e-9bbd-a872ce429218"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 18:04:40 crc kubenswrapper[4777]: I1124 18:04:40.483955 4777 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca02a1d6-3048-418e-9bbd-a872ce429218-host\") on node \"crc\" DevicePath \"\"" Nov 24 18:04:40 crc kubenswrapper[4777]: I1124 18:04:40.494708 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca02a1d6-3048-418e-9bbd-a872ce429218-kube-api-access-2hvxz" (OuterVolumeSpecName: "kube-api-access-2hvxz") pod "ca02a1d6-3048-418e-9bbd-a872ce429218" (UID: "ca02a1d6-3048-418e-9bbd-a872ce429218"). InnerVolumeSpecName "kube-api-access-2hvxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:04:40 crc kubenswrapper[4777]: I1124 18:04:40.586859 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hvxz\" (UniqueName: \"kubernetes.io/projected/ca02a1d6-3048-418e-9bbd-a872ce429218-kube-api-access-2hvxz\") on node \"crc\" DevicePath \"\"" Nov 24 18:04:41 crc kubenswrapper[4777]: I1124 18:04:41.210636 4777 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac409f93dae9c91835b20c3eeeb609fa8cb108b81528e203efb88537d29ea812" Nov 24 18:04:41 crc kubenswrapper[4777]: I1124 18:04:41.210689 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/crc-debug-jjztl" Nov 24 18:04:41 crc kubenswrapper[4777]: I1124 18:04:41.261116 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca02a1d6-3048-418e-9bbd-a872ce429218" path="/var/lib/kubelet/pods/ca02a1d6-3048-418e-9bbd-a872ce429218/volumes" Nov 24 18:04:41 crc kubenswrapper[4777]: I1124 18:04:41.645698 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-w8n5g/crc-debug-ptblp"] Nov 24 18:04:41 crc kubenswrapper[4777]: E1124 18:04:41.646128 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca02a1d6-3048-418e-9bbd-a872ce429218" containerName="container-00" Nov 24 18:04:41 crc kubenswrapper[4777]: I1124 18:04:41.646144 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca02a1d6-3048-418e-9bbd-a872ce429218" containerName="container-00" Nov 24 18:04:41 crc kubenswrapper[4777]: I1124 18:04:41.646369 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca02a1d6-3048-418e-9bbd-a872ce429218" containerName="container-00" Nov 24 18:04:41 crc kubenswrapper[4777]: I1124 18:04:41.647104 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/crc-debug-ptblp" Nov 24 18:04:41 crc kubenswrapper[4777]: I1124 18:04:41.649286 4777 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-w8n5g"/"default-dockercfg-vt6nh" Nov 24 18:04:41 crc kubenswrapper[4777]: I1124 18:04:41.810901 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbsg8\" (UniqueName: \"kubernetes.io/projected/b9632cb2-e240-4d04-a3d1-06576b03e411-kube-api-access-bbsg8\") pod \"crc-debug-ptblp\" (UID: \"b9632cb2-e240-4d04-a3d1-06576b03e411\") " pod="openshift-must-gather-w8n5g/crc-debug-ptblp" Nov 24 18:04:41 crc kubenswrapper[4777]: I1124 18:04:41.811054 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9632cb2-e240-4d04-a3d1-06576b03e411-host\") pod \"crc-debug-ptblp\" (UID: \"b9632cb2-e240-4d04-a3d1-06576b03e411\") " pod="openshift-must-gather-w8n5g/crc-debug-ptblp" Nov 24 18:04:41 crc kubenswrapper[4777]: I1124 18:04:41.912541 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbsg8\" (UniqueName: \"kubernetes.io/projected/b9632cb2-e240-4d04-a3d1-06576b03e411-kube-api-access-bbsg8\") pod \"crc-debug-ptblp\" (UID: \"b9632cb2-e240-4d04-a3d1-06576b03e411\") " pod="openshift-must-gather-w8n5g/crc-debug-ptblp" Nov 24 18:04:41 crc kubenswrapper[4777]: I1124 18:04:41.912644 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9632cb2-e240-4d04-a3d1-06576b03e411-host\") pod \"crc-debug-ptblp\" (UID: \"b9632cb2-e240-4d04-a3d1-06576b03e411\") " pod="openshift-must-gather-w8n5g/crc-debug-ptblp" Nov 24 18:04:41 crc kubenswrapper[4777]: I1124 18:04:41.912738 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9632cb2-e240-4d04-a3d1-06576b03e411-host\") pod \"crc-debug-ptblp\" (UID: \"b9632cb2-e240-4d04-a3d1-06576b03e411\") " pod="openshift-must-gather-w8n5g/crc-debug-ptblp" Nov 24 18:04:41 crc kubenswrapper[4777]: I1124 18:04:41.942708 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbsg8\" (UniqueName: \"kubernetes.io/projected/b9632cb2-e240-4d04-a3d1-06576b03e411-kube-api-access-bbsg8\") pod \"crc-debug-ptblp\" (UID: \"b9632cb2-e240-4d04-a3d1-06576b03e411\") " pod="openshift-must-gather-w8n5g/crc-debug-ptblp" Nov 24 18:04:41 crc kubenswrapper[4777]: I1124 18:04:41.966287 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/crc-debug-ptblp" Nov 24 18:04:42 crc kubenswrapper[4777]: W1124 18:04:42.009930 4777 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9632cb2_e240_4d04_a3d1_06576b03e411.slice/crio-9149e46a55c2c3b0614aee65f4ecc4de2923e2c68569178677cbd00c536bdfe7 WatchSource:0}: Error finding container 9149e46a55c2c3b0614aee65f4ecc4de2923e2c68569178677cbd00c536bdfe7: Status 404 returned error can't find the container with id 9149e46a55c2c3b0614aee65f4ecc4de2923e2c68569178677cbd00c536bdfe7 Nov 24 18:04:42 crc kubenswrapper[4777]: I1124 18:04:42.226824 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w8n5g/crc-debug-ptblp" event={"ID":"b9632cb2-e240-4d04-a3d1-06576b03e411","Type":"ContainerStarted","Data":"9149e46a55c2c3b0614aee65f4ecc4de2923e2c68569178677cbd00c536bdfe7"} Nov 24 18:04:43 crc kubenswrapper[4777]: I1124 18:04:43.247624 4777 generic.go:334] "Generic (PLEG): container finished" podID="b9632cb2-e240-4d04-a3d1-06576b03e411" containerID="17780fda47330e69d2f7d63617c6a3ee524cf6d4f797507705d601c8a93ba6bd" exitCode=0 Nov 24 18:04:43 crc kubenswrapper[4777]: I1124 18:04:43.261755 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w8n5g/crc-debug-ptblp" event={"ID":"b9632cb2-e240-4d04-a3d1-06576b03e411","Type":"ContainerDied","Data":"17780fda47330e69d2f7d63617c6a3ee524cf6d4f797507705d601c8a93ba6bd"} Nov 24 18:04:43 crc kubenswrapper[4777]: I1124 18:04:43.306455 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-w8n5g/crc-debug-ptblp"] Nov 24 18:04:43 crc kubenswrapper[4777]: I1124 18:04:43.313623 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-w8n5g/crc-debug-ptblp"] Nov 24 18:04:44 crc kubenswrapper[4777]: I1124 18:04:44.378112 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/crc-debug-ptblp" Nov 24 18:04:44 crc kubenswrapper[4777]: I1124 18:04:44.468259 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbsg8\" (UniqueName: \"kubernetes.io/projected/b9632cb2-e240-4d04-a3d1-06576b03e411-kube-api-access-bbsg8\") pod \"b9632cb2-e240-4d04-a3d1-06576b03e411\" (UID: \"b9632cb2-e240-4d04-a3d1-06576b03e411\") " Nov 24 18:04:44 crc kubenswrapper[4777]: I1124 18:04:44.468301 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9632cb2-e240-4d04-a3d1-06576b03e411-host\") pod \"b9632cb2-e240-4d04-a3d1-06576b03e411\" (UID: \"b9632cb2-e240-4d04-a3d1-06576b03e411\") " Nov 24 18:04:44 crc kubenswrapper[4777]: I1124 18:04:44.468441 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9632cb2-e240-4d04-a3d1-06576b03e411-host" (OuterVolumeSpecName: "host") pod "b9632cb2-e240-4d04-a3d1-06576b03e411" (UID: "b9632cb2-e240-4d04-a3d1-06576b03e411"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 18:04:44 crc kubenswrapper[4777]: I1124 18:04:44.469017 4777 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9632cb2-e240-4d04-a3d1-06576b03e411-host\") on node \"crc\" DevicePath \"\"" Nov 24 18:04:44 crc kubenswrapper[4777]: I1124 18:04:44.473426 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9632cb2-e240-4d04-a3d1-06576b03e411-kube-api-access-bbsg8" (OuterVolumeSpecName: "kube-api-access-bbsg8") pod "b9632cb2-e240-4d04-a3d1-06576b03e411" (UID: "b9632cb2-e240-4d04-a3d1-06576b03e411"). InnerVolumeSpecName "kube-api-access-bbsg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:04:44 crc kubenswrapper[4777]: I1124 18:04:44.571316 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbsg8\" (UniqueName: \"kubernetes.io/projected/b9632cb2-e240-4d04-a3d1-06576b03e411-kube-api-access-bbsg8\") on node \"crc\" DevicePath \"\"" Nov 24 18:04:45 crc kubenswrapper[4777]: I1124 18:04:45.259571 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9632cb2-e240-4d04-a3d1-06576b03e411" path="/var/lib/kubelet/pods/b9632cb2-e240-4d04-a3d1-06576b03e411/volumes" Nov 24 18:04:45 crc kubenswrapper[4777]: I1124 18:04:45.267980 4777 scope.go:117] "RemoveContainer" containerID="17780fda47330e69d2f7d63617c6a3ee524cf6d4f797507705d601c8a93ba6bd" Nov 24 18:04:45 crc kubenswrapper[4777]: I1124 18:04:45.268123 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/crc-debug-ptblp" Nov 24 18:04:52 crc kubenswrapper[4777]: I1124 18:04:52.245929 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:04:52 crc kubenswrapper[4777]: E1124 18:04:52.246748 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:05:04 crc kubenswrapper[4777]: I1124 18:05:04.245049 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:05:04 crc kubenswrapper[4777]: E1124 18:05:04.245829 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:05:14 crc kubenswrapper[4777]: I1124 18:05:14.065330 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_055503f1-ab1c-47db-be36-37a3f572be32/init-config-reloader/0.log" Nov 24 18:05:14 crc kubenswrapper[4777]: I1124 18:05:14.368522 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_055503f1-ab1c-47db-be36-37a3f572be32/alertmanager/0.log" Nov 24 18:05:14 crc kubenswrapper[4777]: I1124 18:05:14.369258 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_055503f1-ab1c-47db-be36-37a3f572be32/init-config-reloader/0.log" Nov 24 18:05:14 crc kubenswrapper[4777]: I1124 18:05:14.430060 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_055503f1-ab1c-47db-be36-37a3f572be32/config-reloader/0.log" Nov 24 18:05:14 crc kubenswrapper[4777]: I1124 18:05:14.525122 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-65bb697bdb-4pwlb_27a38d66-1471-470a-95e9-c4bfa5ba415b/barbican-api-log/0.log" Nov 24 18:05:14 crc kubenswrapper[4777]: I1124 18:05:14.547341 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-65bb697bdb-4pwlb_27a38d66-1471-470a-95e9-c4bfa5ba415b/barbican-api/0.log" Nov 24 18:05:14 crc kubenswrapper[4777]: I1124 18:05:14.612579 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-845875d46-n8hgg_d173b040-b120-4c77-ba3c-b13c41a6ac37/barbican-keystone-listener/0.log" Nov 24 18:05:14 crc kubenswrapper[4777]: I1124 18:05:14.785275 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5d57df9989-fbq7p_dded6598-74bf-4395-88e0-37152b4dd16c/barbican-worker/0.log" Nov 24 18:05:14 crc kubenswrapper[4777]: I1124 18:05:14.799862 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-845875d46-n8hgg_d173b040-b120-4c77-ba3c-b13c41a6ac37/barbican-keystone-listener-log/0.log" Nov 24 18:05:14 crc kubenswrapper[4777]: I1124 18:05:14.928720 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5d57df9989-fbq7p_dded6598-74bf-4395-88e0-37152b4dd16c/barbican-worker-log/0.log" Nov 24 18:05:15 crc kubenswrapper[4777]: I1124 18:05:15.182162 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-2p42g_1bda09cc-b84e-409f-b4cd-fc387aa02c61/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:15 crc kubenswrapper[4777]: I1124 18:05:15.459648 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f80e34e-82d8-4d61-91b9-323898efdabf/ceilometer-central-agent/0.log" Nov 24 18:05:15 crc kubenswrapper[4777]: I1124 18:05:15.488261 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f80e34e-82d8-4d61-91b9-323898efdabf/proxy-httpd/0.log" Nov 24 18:05:15 crc kubenswrapper[4777]: I1124 18:05:15.512640 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f80e34e-82d8-4d61-91b9-323898efdabf/ceilometer-notification-agent/0.log" Nov 24 18:05:15 crc kubenswrapper[4777]: I1124 18:05:15.606247 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f80e34e-82d8-4d61-91b9-323898efdabf/sg-core/0.log" Nov 24 18:05:15 crc kubenswrapper[4777]: I1124 18:05:15.738261 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_20f43f58-1756-48e1-a958-b49bc82ef1b4/cinder-api-log/0.log" Nov 24 18:05:15 crc kubenswrapper[4777]: I1124 18:05:15.745433 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_20f43f58-1756-48e1-a958-b49bc82ef1b4/cinder-api/0.log" Nov 24 18:05:15 crc kubenswrapper[4777]: I1124 18:05:15.967032 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13/probe/0.log" Nov 24 18:05:15 crc kubenswrapper[4777]: I1124 18:05:15.986161 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f3c3dcf1-2563-4ebb-82a6-f7c995dd9c13/cinder-scheduler/0.log" Nov 24 18:05:16 crc kubenswrapper[4777]: I1124 18:05:16.134455 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_5f4b4358-805a-4a55-a5a8-9200c0c5e5b5/cloudkitty-api/0.log" Nov 24 18:05:16 crc kubenswrapper[4777]: I1124 18:05:16.158480 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_5f4b4358-805a-4a55-a5a8-9200c0c5e5b5/cloudkitty-api-log/0.log" Nov 24 18:05:16 crc kubenswrapper[4777]: I1124 18:05:16.245746 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:05:16 crc kubenswrapper[4777]: E1124 18:05:16.246152 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:05:16 crc kubenswrapper[4777]: I1124 18:05:16.309701 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-compactor-0_e99dec76-e21f-4aab-bdc9-ae098391ee6e/loki-compactor/0.log" Nov 24 18:05:16 crc kubenswrapper[4777]: I1124 18:05:16.448920 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-distributor-56cd74f89f-jg9l8_9c73a207-d21c-4157-b27a-41656a5f6af1/loki-distributor/0.log" Nov 24 18:05:16 crc kubenswrapper[4777]: I1124 18:05:16.701228 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-76cc998948-qskdh_f1facf54-ce8f-441b-bdef-d864d3bc5dd7/gateway/0.log" Nov 24 18:05:16 crc kubenswrapper[4777]: I1124 18:05:16.753915 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-76cc998948-znzk8_1aec5e66-69c2-404c-865a-0995f09fd435/gateway/0.log" Nov 24 18:05:17 crc kubenswrapper[4777]: I1124 18:05:17.021329 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-index-gateway-0_8ddeb3e7-8676-40b5-918b-6d8ac12a2b35/loki-index-gateway/0.log" Nov 24 18:05:17 crc kubenswrapper[4777]: I1124 18:05:17.097703 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-ingester-0_f68bfbdc-e400-488d-9f32-ea90fd1160e3/loki-ingester/0.log" Nov 24 18:05:17 crc kubenswrapper[4777]: I1124 18:05:17.320545 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-query-frontend-779849886d-x7zhc_0d98d291-cc6c-4c5b-90a8-fa2bca9e9f83/loki-query-frontend/0.log" Nov 24 18:05:17 crc kubenswrapper[4777]: I1124 18:05:17.541366 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-querier-548665d79b-tqjct_c0995ddd-f0a2-4137-ac83-2aaae1ce1aeb/loki-querier/0.log" Nov 24 18:05:17 crc kubenswrapper[4777]: I1124 18:05:17.928903 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-sv8lz_45884cb1-7d6c-47a3-b64f-939c1119d70b/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:18 crc kubenswrapper[4777]: I1124 18:05:18.059333 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-qp66f_b21ec255-1398-4613-b461-29fd5ce00ab1/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:18 crc kubenswrapper[4777]: I1124 18:05:18.241230 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c4b758ff5-frg8p_33b50ce8-58e3-457c-81ce-dde13962d338/init/0.log" Nov 24 18:05:18 crc kubenswrapper[4777]: I1124 18:05:18.661894 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c4b758ff5-frg8p_33b50ce8-58e3-457c-81ce-dde13962d338/dnsmasq-dns/0.log" Nov 24 18:05:18 crc kubenswrapper[4777]: I1124 18:05:18.761226 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c4b758ff5-frg8p_33b50ce8-58e3-457c-81ce-dde13962d338/init/0.log" Nov 24 18:05:18 crc kubenswrapper[4777]: I1124 18:05:18.853516 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-9jg8j_266eceac-bfd7-4597-b218-62cea462ce5c/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:19 crc kubenswrapper[4777]: I1124 18:05:19.021615 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_44214c39-ae9a-4e14-a7ec-fc1e95051d7f/glance-httpd/0.log" Nov 24 18:05:19 crc kubenswrapper[4777]: I1124 18:05:19.098563 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-proc-0_1c366c57-2d20-4e76-b90f-c90fc01347fc/cloudkitty-proc/0.log" Nov 24 18:05:19 crc kubenswrapper[4777]: I1124 18:05:19.129923 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_44214c39-ae9a-4e14-a7ec-fc1e95051d7f/glance-log/0.log" Nov 24 18:05:19 crc kubenswrapper[4777]: I1124 18:05:19.242828 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b698d8b2-ca7b-43e7-8da9-aef7d3ba1538/glance-httpd/0.log" Nov 24 18:05:19 crc kubenswrapper[4777]: I1124 18:05:19.275098 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b698d8b2-ca7b-43e7-8da9-aef7d3ba1538/glance-log/0.log" Nov 24 18:05:19 crc kubenswrapper[4777]: I1124 18:05:19.311433 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-4w8nw_0a98c33f-1b9f-4419-8dd3-ea3ba5695980/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:19 crc kubenswrapper[4777]: I1124 18:05:19.494701 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-rvlp6_346142b6-2a63-4610-b1d0-bfebac8c7c61/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:19 crc kubenswrapper[4777]: I1124 18:05:19.628624 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29400121-tzwwt_33e57905-033a-4880-84e8-fe2d92272f5b/keystone-cron/0.log" Nov 24 18:05:19 crc kubenswrapper[4777]: I1124 18:05:19.743724 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_1e1ab5a6-3bb5-4722-9220-1a033a656b93/kube-state-metrics/0.log" Nov 24 18:05:19 crc kubenswrapper[4777]: I1124 18:05:19.805681 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5877679d68-q5nlj_8021c40d-848c-4365-84f2-9878b03d4656/keystone-api/0.log" Nov 24 18:05:19 crc kubenswrapper[4777]: I1124 18:05:19.901175 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-md792_fd79eb9e-a6a2-4932-8a81-27a115ccaef6/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:20 crc kubenswrapper[4777]: I1124 18:05:20.359612 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5984748495-2rxn6_42e06676-5c5b-4684-b525-669d27d31818/neutron-httpd/0.log" Nov 24 18:05:20 crc kubenswrapper[4777]: I1124 18:05:20.447530 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5984748495-2rxn6_42e06676-5c5b-4684-b525-669d27d31818/neutron-api/0.log" Nov 24 18:05:20 crc kubenswrapper[4777]: I1124 18:05:20.469308 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-jrjl2_97805347-c06c-409d-a299-cd4cfdf7e98e/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:20 crc kubenswrapper[4777]: I1124 18:05:20.997725 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_f474cf0f-37fd-495d-ac31-493641c222a7/nova-api-log/0.log" Nov 24 18:05:21 crc kubenswrapper[4777]: I1124 18:05:21.156854 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_aee763a5-b286-4fdd-9573-d06a0e915f9e/nova-cell0-conductor-conductor/0.log" Nov 24 18:05:21 crc kubenswrapper[4777]: I1124 18:05:21.488632 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_f474cf0f-37fd-495d-ac31-493641c222a7/nova-api-api/0.log" Nov 24 18:05:21 crc kubenswrapper[4777]: I1124 18:05:21.547612 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b929fcfa-b979-4f40-a469-fdd1e24a9ef2/nova-cell1-conductor-conductor/0.log" Nov 24 18:05:21 crc kubenswrapper[4777]: I1124 18:05:21.573508 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_ad8c715b-ea28-4366-882c-4371692973f0/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 18:05:21 crc kubenswrapper[4777]: I1124 18:05:21.757045 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-w44w5_9651259d-da1a-4580-855a-94a9310fc952/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:21 crc kubenswrapper[4777]: I1124 18:05:21.870302 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4da8a01e-98cb-4d18-b86e-c281010dcf8e/nova-metadata-log/0.log" Nov 24 18:05:22 crc kubenswrapper[4777]: I1124 18:05:22.564154 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4eba3f22-b7ad-464d-a0c9-437c449d124b/mysql-bootstrap/0.log" Nov 24 18:05:22 crc kubenswrapper[4777]: I1124 18:05:22.566609 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_2a63c908-3ee5-449c-ab1c-6ca209338bee/nova-scheduler-scheduler/0.log" Nov 24 18:05:22 crc kubenswrapper[4777]: I1124 18:05:22.795395 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4eba3f22-b7ad-464d-a0c9-437c449d124b/galera/0.log" Nov 24 18:05:22 crc kubenswrapper[4777]: I1124 18:05:22.847788 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4eba3f22-b7ad-464d-a0c9-437c449d124b/mysql-bootstrap/0.log" Nov 24 18:05:23 crc kubenswrapper[4777]: I1124 18:05:23.023539 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1d71a086-37e0-4409-9db9-66e6f1374406/mysql-bootstrap/0.log" Nov 24 18:05:23 crc kubenswrapper[4777]: I1124 18:05:23.211481 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1d71a086-37e0-4409-9db9-66e6f1374406/mysql-bootstrap/0.log" Nov 24 18:05:23 crc kubenswrapper[4777]: I1124 18:05:23.217215 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1d71a086-37e0-4409-9db9-66e6f1374406/galera/0.log" Nov 24 18:05:23 crc kubenswrapper[4777]: I1124 18:05:23.409220 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4da8a01e-98cb-4d18-b86e-c281010dcf8e/nova-metadata-metadata/0.log" Nov 24 18:05:23 crc kubenswrapper[4777]: I1124 18:05:23.477844 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2d497520-8510-44d4-b57c-9b8e35ebaa32/openstackclient/0.log" Nov 24 18:05:23 crc kubenswrapper[4777]: I1124 18:05:23.573821 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-lgxzj_04bed066-31b6-49d0-90ee-68e38f7944c1/openstack-network-exporter/0.log" Nov 24 18:05:23 crc kubenswrapper[4777]: I1124 18:05:23.793761 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dkbds_3641f728-c179-4f10-93c2-98f74763e77f/ovsdb-server-init/0.log" Nov 24 18:05:24 crc kubenswrapper[4777]: I1124 18:05:24.045148 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dkbds_3641f728-c179-4f10-93c2-98f74763e77f/ovsdb-server/0.log" Nov 24 18:05:24 crc kubenswrapper[4777]: I1124 18:05:24.046315 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dkbds_3641f728-c179-4f10-93c2-98f74763e77f/ovsdb-server-init/0.log" Nov 24 18:05:24 crc kubenswrapper[4777]: I1124 18:05:24.068557 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dkbds_3641f728-c179-4f10-93c2-98f74763e77f/ovs-vswitchd/0.log" Nov 24 18:05:24 crc kubenswrapper[4777]: I1124 18:05:24.269424 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-q4l2k_73928e73-667c-4b69-aed9-72f32012fbdc/ovn-controller/0.log" Nov 24 18:05:24 crc kubenswrapper[4777]: I1124 18:05:24.303879 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-xkd2c_9c522429-96d7-44c9-812c-f86726228305/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:24 crc kubenswrapper[4777]: I1124 18:05:24.520266 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_001aba35-921b-43b5-87e4-7e92d5595a7b/openstack-network-exporter/0.log" Nov 24 18:05:24 crc kubenswrapper[4777]: I1124 18:05:24.618259 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_001aba35-921b-43b5-87e4-7e92d5595a7b/ovn-northd/0.log" Nov 24 18:05:24 crc kubenswrapper[4777]: I1124 18:05:24.871522 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4ccca46a-1cac-4693-8490-2612b138c9d5/ovsdbserver-nb/0.log" Nov 24 18:05:24 crc kubenswrapper[4777]: I1124 18:05:24.931608 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4ccca46a-1cac-4693-8490-2612b138c9d5/openstack-network-exporter/0.log" Nov 24 18:05:24 crc kubenswrapper[4777]: I1124 18:05:24.944526 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_737d5505-10bc-4e56-b846-01a3af071b38/openstack-network-exporter/0.log" Nov 24 18:05:25 crc kubenswrapper[4777]: I1124 18:05:25.123596 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_737d5505-10bc-4e56-b846-01a3af071b38/ovsdbserver-sb/0.log" Nov 24 18:05:25 crc kubenswrapper[4777]: I1124 18:05:25.295775 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-55f98d7b-5njfr_b25fa9b0-aec5-4a33-aefe-2ee5685d0e88/placement-api/0.log" Nov 24 18:05:25 crc kubenswrapper[4777]: I1124 18:05:25.387748 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-55f98d7b-5njfr_b25fa9b0-aec5-4a33-aefe-2ee5685d0e88/placement-log/0.log" Nov 24 18:05:25 crc kubenswrapper[4777]: I1124 18:05:25.429955 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_df409568-e0d5-49db-8da4-efbe45ec3a2c/init-config-reloader/0.log" Nov 24 18:05:25 crc kubenswrapper[4777]: I1124 18:05:25.586130 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_df409568-e0d5-49db-8da4-efbe45ec3a2c/init-config-reloader/0.log" Nov 24 18:05:25 crc kubenswrapper[4777]: I1124 18:05:25.694298 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_df409568-e0d5-49db-8da4-efbe45ec3a2c/thanos-sidecar/0.log" Nov 24 18:05:25 crc kubenswrapper[4777]: I1124 18:05:25.735375 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_df409568-e0d5-49db-8da4-efbe45ec3a2c/config-reloader/0.log" Nov 24 18:05:25 crc kubenswrapper[4777]: I1124 18:05:25.771379 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_df409568-e0d5-49db-8da4-efbe45ec3a2c/prometheus/0.log" Nov 24 18:05:26 crc kubenswrapper[4777]: I1124 18:05:26.251204 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_42d2ab97-ca44-4234-8405-22f8797e0c0b/setup-container/0.log" Nov 24 18:05:26 crc kubenswrapper[4777]: I1124 18:05:26.436707 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_42d2ab97-ca44-4234-8405-22f8797e0c0b/setup-container/0.log" Nov 24 18:05:26 crc kubenswrapper[4777]: I1124 18:05:26.505670 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_42d2ab97-ca44-4234-8405-22f8797e0c0b/rabbitmq/0.log" Nov 24 18:05:26 crc kubenswrapper[4777]: I1124 18:05:26.571085 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_78712b89-27c4-4481-b152-320fb7fee9fc/setup-container/0.log" Nov 24 18:05:27 crc kubenswrapper[4777]: I1124 18:05:27.428450 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:05:27 crc kubenswrapper[4777]: I1124 18:05:27.435051 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_78712b89-27c4-4481-b152-320fb7fee9fc/setup-container/0.log" Nov 24 18:05:27 crc kubenswrapper[4777]: E1124 18:05:27.461172 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:05:27 crc kubenswrapper[4777]: I1124 18:05:27.483007 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-8qwj6_902e59e0-e5a3-4f96-953e-ec961dbe53d9/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:27 crc kubenswrapper[4777]: I1124 18:05:27.553190 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_78712b89-27c4-4481-b152-320fb7fee9fc/rabbitmq/0.log" Nov 24 18:05:27 crc kubenswrapper[4777]: I1124 18:05:27.765705 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-qfmq5_f27aca44-4a34-469c-97fd-075eed1300dd/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:27 crc kubenswrapper[4777]: I1124 18:05:27.903644 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-6nqtq_10de4ea3-2dc3-4e57-ac55-d6be12743347/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:28 crc kubenswrapper[4777]: I1124 18:05:28.102239 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-kq565_a57c6a80-0881-430c-a6bb-beea385e0417/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:28 crc kubenswrapper[4777]: I1124 18:05:28.160636 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-q4nrb_3be75c8d-7695-403e-b801-d2650e0b331c/ssh-known-hosts-edpm-deployment/0.log" Nov 24 18:05:28 crc kubenswrapper[4777]: I1124 18:05:28.412794 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-644f75df95-wzmtb_22f2176a-dcae-476f-92b5-eb3d19d9bb4f/proxy-server/0.log" Nov 24 18:05:28 crc kubenswrapper[4777]: I1124 18:05:28.479991 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-xgm6s_3ec24d0c-556b-4a24-8f42-38fed08efe6b/swift-ring-rebalance/0.log" Nov 24 18:05:28 crc kubenswrapper[4777]: I1124 18:05:28.481404 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-644f75df95-wzmtb_22f2176a-dcae-476f-92b5-eb3d19d9bb4f/proxy-httpd/0.log" Nov 24 18:05:28 crc kubenswrapper[4777]: I1124 18:05:28.681566 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/account-replicator/0.log" Nov 24 18:05:28 crc kubenswrapper[4777]: I1124 18:05:28.710861 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/account-auditor/0.log" Nov 24 18:05:28 crc kubenswrapper[4777]: I1124 18:05:28.788094 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/account-reaper/0.log" Nov 24 18:05:29 crc kubenswrapper[4777]: I1124 18:05:29.056650 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/account-server/0.log" Nov 24 18:05:29 crc kubenswrapper[4777]: I1124 18:05:29.072987 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/container-server/0.log" Nov 24 18:05:29 crc kubenswrapper[4777]: I1124 18:05:29.090097 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/container-replicator/0.log" Nov 24 18:05:29 crc kubenswrapper[4777]: I1124 18:05:29.103336 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/container-auditor/0.log" Nov 24 18:05:29 crc kubenswrapper[4777]: I1124 18:05:29.261502 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/container-updater/0.log" Nov 24 18:05:29 crc kubenswrapper[4777]: I1124 18:05:29.292052 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/object-auditor/0.log" Nov 24 18:05:29 crc kubenswrapper[4777]: I1124 18:05:29.326767 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/object-expirer/0.log" Nov 24 18:05:29 crc kubenswrapper[4777]: I1124 18:05:29.371464 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/object-replicator/0.log" Nov 24 18:05:29 crc kubenswrapper[4777]: I1124 18:05:29.514152 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/object-server/0.log" Nov 24 18:05:29 crc kubenswrapper[4777]: I1124 18:05:29.524534 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/rsync/0.log" Nov 24 18:05:29 crc kubenswrapper[4777]: I1124 18:05:29.590564 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/object-updater/0.log" Nov 24 18:05:29 crc kubenswrapper[4777]: I1124 18:05:29.617412 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9c412d58-6198-4254-932d-5eb90420f8ba/swift-recon-cron/0.log" Nov 24 18:05:29 crc kubenswrapper[4777]: I1124 18:05:29.965232 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-hqqgl_bcd7d063-a7cc-4f12-8ba2-9e7578152a3e/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:30 crc kubenswrapper[4777]: I1124 18:05:30.204944 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_08d58d9b-d4f1-458d-bfe3-fa8bc585f352/tempest-tests-tempest-tests-runner/0.log" Nov 24 18:05:30 crc kubenswrapper[4777]: I1124 18:05:30.353283 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_2d149ad6-8d86-445c-baed-c43f1eaec2cb/test-operator-logs-container/0.log" Nov 24 18:05:30 crc kubenswrapper[4777]: I1124 18:05:30.463443 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-c2w8j_541aa87c-3013-4641-950b-a23f825bcb96/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 18:05:35 crc kubenswrapper[4777]: I1124 18:05:35.535645 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_0254e0e3-f9b9-4017-954a-14d3d21e37c5/memcached/0.log" Nov 24 18:05:42 crc kubenswrapper[4777]: I1124 18:05:42.246626 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:05:42 crc kubenswrapper[4777]: E1124 18:05:42.247360 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:05:56 crc kubenswrapper[4777]: I1124 18:05:56.245694 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:05:56 crc kubenswrapper[4777]: I1124 18:05:56.971006 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"1b72042dc72b04f0b4e64de35ea4d6163e72725033a38990bed0f8f82baa07f0"} Nov 24 18:05:58 crc kubenswrapper[4777]: I1124 18:05:58.674679 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl_bd350075-1ebb-4431-9b02-91e2d6c5e7d3/util/0.log" Nov 24 18:05:59 crc kubenswrapper[4777]: I1124 18:05:59.011276 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl_bd350075-1ebb-4431-9b02-91e2d6c5e7d3/pull/0.log" Nov 24 18:05:59 crc kubenswrapper[4777]: I1124 18:05:59.013061 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl_bd350075-1ebb-4431-9b02-91e2d6c5e7d3/util/0.log" Nov 24 18:05:59 crc kubenswrapper[4777]: I1124 18:05:59.056484 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl_bd350075-1ebb-4431-9b02-91e2d6c5e7d3/pull/0.log" Nov 24 18:05:59 crc kubenswrapper[4777]: I1124 18:05:59.326731 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl_bd350075-1ebb-4431-9b02-91e2d6c5e7d3/util/0.log" Nov 24 18:05:59 crc kubenswrapper[4777]: I1124 18:05:59.331922 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl_bd350075-1ebb-4431-9b02-91e2d6c5e7d3/extract/0.log" Nov 24 18:05:59 crc kubenswrapper[4777]: I1124 18:05:59.408260 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_487e9e2af153ab3122f10a5439116d2fbe109130520c2846c8a37103cepg8tl_bd350075-1ebb-4431-9b02-91e2d6c5e7d3/pull/0.log" Nov 24 18:05:59 crc kubenswrapper[4777]: I1124 18:05:59.747298 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-82ppc_a7457835-23e4-460e-a155-07c1d4e7c30e/kube-rbac-proxy/0.log" Nov 24 18:05:59 crc kubenswrapper[4777]: I1124 18:05:59.769900 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-82ppc_a7457835-23e4-460e-a155-07c1d4e7c30e/manager/0.log" Nov 24 18:05:59 crc kubenswrapper[4777]: I1124 18:05:59.906722 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-2chns_1d2ecc53-902f-41f3-80f4-93f9e48c8532/kube-rbac-proxy/0.log" Nov 24 18:06:00 crc kubenswrapper[4777]: I1124 18:06:00.029878 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-2chns_1d2ecc53-902f-41f3-80f4-93f9e48c8532/manager/0.log" Nov 24 18:06:00 crc kubenswrapper[4777]: I1124 18:06:00.091798 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-klpxr_99bfc986-b80d-4193-bf32-4488c11f0066/kube-rbac-proxy/0.log" Nov 24 18:06:00 crc kubenswrapper[4777]: I1124 18:06:00.141416 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-klpxr_99bfc986-b80d-4193-bf32-4488c11f0066/manager/0.log" Nov 24 18:06:00 crc kubenswrapper[4777]: I1124 18:06:00.295098 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-9mzvw_fe817743-7f1e-46a4-9757-5b678ddc097a/kube-rbac-proxy/0.log" Nov 24 18:06:00 crc kubenswrapper[4777]: I1124 18:06:00.383799 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-9mzvw_fe817743-7f1e-46a4-9757-5b678ddc097a/manager/0.log" Nov 24 18:06:00 crc kubenswrapper[4777]: I1124 18:06:00.461156 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-zmms9_1d3fc8a3-f3cf-4f37-846d-b244a52415f8/kube-rbac-proxy/0.log" Nov 24 18:06:00 crc kubenswrapper[4777]: I1124 18:06:00.571904 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-zmms9_1d3fc8a3-f3cf-4f37-846d-b244a52415f8/manager/0.log" Nov 24 18:06:00 crc kubenswrapper[4777]: I1124 18:06:00.635129 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-6xxk9_8401afce-ddb4-4195-b6ae-7e5c91128525/kube-rbac-proxy/0.log" Nov 24 18:06:00 crc kubenswrapper[4777]: I1124 18:06:00.706070 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-6xxk9_8401afce-ddb4-4195-b6ae-7e5c91128525/manager/0.log" Nov 24 18:06:00 crc kubenswrapper[4777]: I1124 18:06:00.803555 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-r5pxc_b755cc8b-dd00-4b28-8fb4-908e0e3db9a5/kube-rbac-proxy/0.log" Nov 24 18:06:00 crc kubenswrapper[4777]: I1124 18:06:00.972578 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-wmvls_f7f44c61-027a-41c4-9665-8394d579ba33/kube-rbac-proxy/0.log" Nov 24 18:06:00 crc kubenswrapper[4777]: I1124 18:06:00.986765 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-r5pxc_b755cc8b-dd00-4b28-8fb4-908e0e3db9a5/manager/0.log" Nov 24 18:06:01 crc kubenswrapper[4777]: I1124 18:06:01.148111 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-wmvls_f7f44c61-027a-41c4-9665-8394d579ba33/manager/0.log" Nov 24 18:06:01 crc kubenswrapper[4777]: I1124 18:06:01.297626 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-v5vgx_f1425713-2fd6-4108-abe3-9f46063e98f5/manager/0.log" Nov 24 18:06:01 crc kubenswrapper[4777]: I1124 18:06:01.300954 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-v5vgx_f1425713-2fd6-4108-abe3-9f46063e98f5/kube-rbac-proxy/0.log" Nov 24 18:06:01 crc kubenswrapper[4777]: I1124 18:06:01.512989 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-mf8r8_38f287c7-6c43-4dc9-b547-0d494704874a/kube-rbac-proxy/0.log" Nov 24 18:06:01 crc kubenswrapper[4777]: I1124 18:06:01.567541 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-mf8r8_38f287c7-6c43-4dc9-b547-0d494704874a/manager/0.log" Nov 24 18:06:01 crc kubenswrapper[4777]: I1124 18:06:01.699508 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-g9kpn_73a88858-0760-4b40-b57f-71e4f9977129/kube-rbac-proxy/0.log" Nov 24 18:06:01 crc kubenswrapper[4777]: I1124 18:06:01.824175 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-g9kpn_73a88858-0760-4b40-b57f-71e4f9977129/manager/0.log" Nov 24 18:06:01 crc kubenswrapper[4777]: I1124 18:06:01.836325 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-dc68g_fd1db7a5-22e3-4f16-a026-0b8895ef292d/kube-rbac-proxy/0.log" Nov 24 18:06:01 crc kubenswrapper[4777]: I1124 18:06:01.959658 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-dc68g_fd1db7a5-22e3-4f16-a026-0b8895ef292d/manager/0.log" Nov 24 18:06:02 crc kubenswrapper[4777]: I1124 18:06:02.135195 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-58vj4_a9a5349a-868d-4c3e-a3ab-f57b55643759/kube-rbac-proxy/0.log" Nov 24 18:06:02 crc kubenswrapper[4777]: I1124 18:06:02.162420 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-58vj4_a9a5349a-868d-4c3e-a3ab-f57b55643759/manager/0.log" Nov 24 18:06:02 crc kubenswrapper[4777]: I1124 18:06:02.449240 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-ss948_7ea09c1e-d85f-435e-a029-8b4f0df06839/kube-rbac-proxy/0.log" Nov 24 18:06:02 crc kubenswrapper[4777]: I1124 18:06:02.490698 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-ss948_7ea09c1e-d85f-435e-a029-8b4f0df06839/manager/0.log" Nov 24 18:06:02 crc kubenswrapper[4777]: I1124 18:06:02.695485 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d_4a491c1c-85db-4efd-b944-40b1651a3c18/kube-rbac-proxy/0.log" Nov 24 18:06:02 crc kubenswrapper[4777]: I1124 18:06:02.797818 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-xzq9d_4a491c1c-85db-4efd-b944-40b1651a3c18/manager/0.log" Nov 24 18:06:02 crc kubenswrapper[4777]: I1124 18:06:02.989012 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7856bf9fdc-k8wq8_783eef36-329f-4ceb-b516-020fe4cf5151/kube-rbac-proxy/0.log" Nov 24 18:06:03 crc kubenswrapper[4777]: I1124 18:06:03.056623 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-68bb788c6f-ft4rz_547759be-2d04-4553-aedc-e3f1b7ef9779/kube-rbac-proxy/0.log" Nov 24 18:06:03 crc kubenswrapper[4777]: I1124 18:06:03.262092 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zvxss_127dc4f2-329f-453e-8f21-1877225cec81/registry-server/0.log" Nov 24 18:06:03 crc kubenswrapper[4777]: I1124 18:06:03.373114 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-68bb788c6f-ft4rz_547759be-2d04-4553-aedc-e3f1b7ef9779/operator/0.log" Nov 24 18:06:03 crc kubenswrapper[4777]: I1124 18:06:03.606603 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-vxgn2_cfd203de-6a3e-4fdf-af55-3f5eaa62c5c7/kube-rbac-proxy/0.log" Nov 24 18:06:03 crc kubenswrapper[4777]: I1124 18:06:03.633524 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-hmsp7_ef1391a6-60c0-47c9-aa4f-7fa39f90d55b/kube-rbac-proxy/0.log" Nov 24 18:06:03 crc kubenswrapper[4777]: I1124 18:06:03.722546 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-vxgn2_cfd203de-6a3e-4fdf-af55-3f5eaa62c5c7/manager/0.log" Nov 24 18:06:04 crc kubenswrapper[4777]: I1124 18:06:04.037045 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-hmsp7_ef1391a6-60c0-47c9-aa4f-7fa39f90d55b/manager/0.log" Nov 24 18:06:04 crc kubenswrapper[4777]: I1124 18:06:04.138542 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-s4ggm_386e5031-5136-4c3f-9ba2-6051496e0ed9/operator/0.log" Nov 24 18:06:04 crc kubenswrapper[4777]: I1124 18:06:04.280280 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-q2p2n_c847a619-1224-4a3c-bc57-bb75924d8553/kube-rbac-proxy/0.log" Nov 24 18:06:04 crc kubenswrapper[4777]: I1124 18:06:04.363295 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-64dc7b4899-mmn8p_6f802034-e827-4870-9e32-d1d0cc263587/kube-rbac-proxy/0.log" Nov 24 18:06:04 crc kubenswrapper[4777]: I1124 18:06:04.474078 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-q2p2n_c847a619-1224-4a3c-bc57-bb75924d8553/manager/0.log" Nov 24 18:06:04 crc kubenswrapper[4777]: I1124 18:06:04.568752 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-jbqdt_2f5425e7-4f78-4b35-bb3f-b1c0065cfc42/kube-rbac-proxy/0.log" Nov 24 18:06:04 crc kubenswrapper[4777]: I1124 18:06:04.743606 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-jbqdt_2f5425e7-4f78-4b35-bb3f-b1c0065cfc42/manager/0.log" Nov 24 18:06:04 crc kubenswrapper[4777]: I1124 18:06:04.822774 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7856bf9fdc-k8wq8_783eef36-329f-4ceb-b516-020fe4cf5151/manager/0.log" Nov 24 18:06:04 crc kubenswrapper[4777]: I1124 18:06:04.872945 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-g5jdq_44e68f24-0747-476d-b3e5-11e2952c3792/kube-rbac-proxy/0.log" Nov 24 18:06:04 crc kubenswrapper[4777]: I1124 18:06:04.985361 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-g5jdq_44e68f24-0747-476d-b3e5-11e2952c3792/manager/0.log" Nov 24 18:06:05 crc kubenswrapper[4777]: I1124 18:06:05.167374 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-64dc7b4899-mmn8p_6f802034-e827-4870-9e32-d1d0cc263587/manager/0.log" Nov 24 18:06:19 crc kubenswrapper[4777]: I1124 18:06:19.669906 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bcmr5"] Nov 24 18:06:19 crc kubenswrapper[4777]: E1124 18:06:19.671048 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9632cb2-e240-4d04-a3d1-06576b03e411" containerName="container-00" Nov 24 18:06:19 crc kubenswrapper[4777]: I1124 18:06:19.671068 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9632cb2-e240-4d04-a3d1-06576b03e411" containerName="container-00" Nov 24 18:06:19 crc kubenswrapper[4777]: I1124 18:06:19.671346 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9632cb2-e240-4d04-a3d1-06576b03e411" containerName="container-00" Nov 24 18:06:19 crc kubenswrapper[4777]: I1124 18:06:19.673329 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:19 crc kubenswrapper[4777]: I1124 18:06:19.693854 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bcmr5"] Nov 24 18:06:19 crc kubenswrapper[4777]: I1124 18:06:19.724164 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295dc21d-5d67-4ae7-a0b5-676cc0366501-catalog-content\") pod \"certified-operators-bcmr5\" (UID: \"295dc21d-5d67-4ae7-a0b5-676cc0366501\") " pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:19 crc kubenswrapper[4777]: I1124 18:06:19.724341 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wj8q\" (UniqueName: \"kubernetes.io/projected/295dc21d-5d67-4ae7-a0b5-676cc0366501-kube-api-access-2wj8q\") pod \"certified-operators-bcmr5\" (UID: \"295dc21d-5d67-4ae7-a0b5-676cc0366501\") " pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:19 crc kubenswrapper[4777]: I1124 18:06:19.724500 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295dc21d-5d67-4ae7-a0b5-676cc0366501-utilities\") pod \"certified-operators-bcmr5\" (UID: \"295dc21d-5d67-4ae7-a0b5-676cc0366501\") " pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:19 crc kubenswrapper[4777]: I1124 18:06:19.826128 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wj8q\" (UniqueName: \"kubernetes.io/projected/295dc21d-5d67-4ae7-a0b5-676cc0366501-kube-api-access-2wj8q\") pod \"certified-operators-bcmr5\" (UID: \"295dc21d-5d67-4ae7-a0b5-676cc0366501\") " pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:19 crc kubenswrapper[4777]: I1124 18:06:19.826249 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295dc21d-5d67-4ae7-a0b5-676cc0366501-utilities\") pod \"certified-operators-bcmr5\" (UID: \"295dc21d-5d67-4ae7-a0b5-676cc0366501\") " pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:19 crc kubenswrapper[4777]: I1124 18:06:19.826324 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295dc21d-5d67-4ae7-a0b5-676cc0366501-catalog-content\") pod \"certified-operators-bcmr5\" (UID: \"295dc21d-5d67-4ae7-a0b5-676cc0366501\") " pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:19 crc kubenswrapper[4777]: I1124 18:06:19.827092 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295dc21d-5d67-4ae7-a0b5-676cc0366501-utilities\") pod \"certified-operators-bcmr5\" (UID: \"295dc21d-5d67-4ae7-a0b5-676cc0366501\") " pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:19 crc kubenswrapper[4777]: I1124 18:06:19.827144 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295dc21d-5d67-4ae7-a0b5-676cc0366501-catalog-content\") pod \"certified-operators-bcmr5\" (UID: \"295dc21d-5d67-4ae7-a0b5-676cc0366501\") " pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:19 crc kubenswrapper[4777]: I1124 18:06:19.844162 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wj8q\" (UniqueName: \"kubernetes.io/projected/295dc21d-5d67-4ae7-a0b5-676cc0366501-kube-api-access-2wj8q\") pod \"certified-operators-bcmr5\" (UID: \"295dc21d-5d67-4ae7-a0b5-676cc0366501\") " pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:19 crc kubenswrapper[4777]: I1124 18:06:19.997864 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:20 crc kubenswrapper[4777]: I1124 18:06:20.566153 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bcmr5"] Nov 24 18:06:21 crc kubenswrapper[4777]: I1124 18:06:21.245846 4777 generic.go:334] "Generic (PLEG): container finished" podID="295dc21d-5d67-4ae7-a0b5-676cc0366501" containerID="42235ab3aac7de732755d099a9c472bd4a09ce3be176a72bc11774cf547635f4" exitCode=0 Nov 24 18:06:21 crc kubenswrapper[4777]: I1124 18:06:21.260508 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bcmr5" event={"ID":"295dc21d-5d67-4ae7-a0b5-676cc0366501","Type":"ContainerDied","Data":"42235ab3aac7de732755d099a9c472bd4a09ce3be176a72bc11774cf547635f4"} Nov 24 18:06:21 crc kubenswrapper[4777]: I1124 18:06:21.260558 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bcmr5" event={"ID":"295dc21d-5d67-4ae7-a0b5-676cc0366501","Type":"ContainerStarted","Data":"67c56b73f64a921cf480b9392e40754884c11d99fad6f3bcbf0df197aa882722"} Nov 24 18:06:22 crc kubenswrapper[4777]: I1124 18:06:22.609835 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zmdsl_35059f02-39e8-40b0-8295-8230591cd996/control-plane-machine-set-operator/0.log" Nov 24 18:06:22 crc kubenswrapper[4777]: I1124 18:06:22.790177 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xtvn9_92efc884-606f-4fca-9410-4856b39a985d/kube-rbac-proxy/0.log" Nov 24 18:06:22 crc kubenswrapper[4777]: I1124 18:06:22.861352 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xtvn9_92efc884-606f-4fca-9410-4856b39a985d/machine-api-operator/0.log" Nov 24 18:06:24 crc kubenswrapper[4777]: I1124 18:06:24.273951 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bcmr5" event={"ID":"295dc21d-5d67-4ae7-a0b5-676cc0366501","Type":"ContainerStarted","Data":"2a50451746d0cdf6b541927860e995d89bfe58c29b2f751594677a7fd23ed797"} Nov 24 18:06:27 crc kubenswrapper[4777]: I1124 18:06:27.988907 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-544hh"] Nov 24 18:06:27 crc kubenswrapper[4777]: I1124 18:06:27.991930 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:06:28 crc kubenswrapper[4777]: I1124 18:06:28.001357 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqqvd\" (UniqueName: \"kubernetes.io/projected/d2135c03-e50b-4f82-807b-82722514714b-kube-api-access-fqqvd\") pod \"redhat-operators-544hh\" (UID: \"d2135c03-e50b-4f82-807b-82722514714b\") " pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:06:28 crc kubenswrapper[4777]: I1124 18:06:28.001402 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2135c03-e50b-4f82-807b-82722514714b-catalog-content\") pod \"redhat-operators-544hh\" (UID: \"d2135c03-e50b-4f82-807b-82722514714b\") " pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:06:28 crc kubenswrapper[4777]: I1124 18:06:28.001652 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2135c03-e50b-4f82-807b-82722514714b-utilities\") pod \"redhat-operators-544hh\" (UID: \"d2135c03-e50b-4f82-807b-82722514714b\") " pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:06:28 crc kubenswrapper[4777]: I1124 18:06:28.011559 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-544hh"] Nov 24 18:06:28 crc kubenswrapper[4777]: I1124 18:06:28.103568 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2135c03-e50b-4f82-807b-82722514714b-utilities\") pod \"redhat-operators-544hh\" (UID: \"d2135c03-e50b-4f82-807b-82722514714b\") " pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:06:28 crc kubenswrapper[4777]: I1124 18:06:28.103921 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqqvd\" (UniqueName: \"kubernetes.io/projected/d2135c03-e50b-4f82-807b-82722514714b-kube-api-access-fqqvd\") pod \"redhat-operators-544hh\" (UID: \"d2135c03-e50b-4f82-807b-82722514714b\") " pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:06:28 crc kubenswrapper[4777]: I1124 18:06:28.103941 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2135c03-e50b-4f82-807b-82722514714b-catalog-content\") pod \"redhat-operators-544hh\" (UID: \"d2135c03-e50b-4f82-807b-82722514714b\") " pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:06:28 crc kubenswrapper[4777]: I1124 18:06:28.104748 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2135c03-e50b-4f82-807b-82722514714b-utilities\") pod \"redhat-operators-544hh\" (UID: \"d2135c03-e50b-4f82-807b-82722514714b\") " pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:06:28 crc kubenswrapper[4777]: I1124 18:06:28.105192 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2135c03-e50b-4f82-807b-82722514714b-catalog-content\") pod \"redhat-operators-544hh\" (UID: \"d2135c03-e50b-4f82-807b-82722514714b\") " pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:06:28 crc kubenswrapper[4777]: I1124 18:06:28.137626 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqqvd\" (UniqueName: \"kubernetes.io/projected/d2135c03-e50b-4f82-807b-82722514714b-kube-api-access-fqqvd\") pod \"redhat-operators-544hh\" (UID: \"d2135c03-e50b-4f82-807b-82722514714b\") " pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:06:28 crc kubenswrapper[4777]: I1124 18:06:28.310630 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:06:28 crc kubenswrapper[4777]: I1124 18:06:28.837284 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-544hh"] Nov 24 18:06:29 crc kubenswrapper[4777]: I1124 18:06:29.339255 4777 generic.go:334] "Generic (PLEG): container finished" podID="d2135c03-e50b-4f82-807b-82722514714b" containerID="2cc12217b1faa6c8a5b5f12c83bc3a703f9284a9f5f11da9bb7e06200756e2ea" exitCode=0 Nov 24 18:06:29 crc kubenswrapper[4777]: I1124 18:06:29.339451 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-544hh" event={"ID":"d2135c03-e50b-4f82-807b-82722514714b","Type":"ContainerDied","Data":"2cc12217b1faa6c8a5b5f12c83bc3a703f9284a9f5f11da9bb7e06200756e2ea"} Nov 24 18:06:29 crc kubenswrapper[4777]: I1124 18:06:29.339597 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-544hh" event={"ID":"d2135c03-e50b-4f82-807b-82722514714b","Type":"ContainerStarted","Data":"3da45c2ea443b2d94753150099821ef97dc570d02d622905358640abf7d80f59"} Nov 24 18:06:29 crc kubenswrapper[4777]: I1124 18:06:29.341680 4777 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 18:06:29 crc kubenswrapper[4777]: I1124 18:06:29.342323 4777 generic.go:334] "Generic (PLEG): container finished" podID="295dc21d-5d67-4ae7-a0b5-676cc0366501" containerID="2a50451746d0cdf6b541927860e995d89bfe58c29b2f751594677a7fd23ed797" exitCode=0 Nov 24 18:06:29 crc kubenswrapper[4777]: I1124 18:06:29.342352 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bcmr5" event={"ID":"295dc21d-5d67-4ae7-a0b5-676cc0366501","Type":"ContainerDied","Data":"2a50451746d0cdf6b541927860e995d89bfe58c29b2f751594677a7fd23ed797"} Nov 24 18:06:30 crc kubenswrapper[4777]: I1124 18:06:30.355132 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bcmr5" event={"ID":"295dc21d-5d67-4ae7-a0b5-676cc0366501","Type":"ContainerStarted","Data":"506307466a6cb972e7b9ed3400e0f9a0707ee32f5d99fc7bbf6c2acb68c26c75"} Nov 24 18:06:30 crc kubenswrapper[4777]: I1124 18:06:30.387372 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bcmr5" podStartSLOduration=2.606989307 podStartE2EDuration="11.38733721s" podCreationTimestamp="2025-11-24 18:06:19 +0000 UTC" firstStartedPulling="2025-11-24 18:06:21.250161103 +0000 UTC m=+3939.409356152" lastFinishedPulling="2025-11-24 18:06:30.030509006 +0000 UTC m=+3948.189704055" observedRunningTime="2025-11-24 18:06:30.377801874 +0000 UTC m=+3948.536996943" watchObservedRunningTime="2025-11-24 18:06:30.38733721 +0000 UTC m=+3948.546532259" Nov 24 18:06:31 crc kubenswrapper[4777]: I1124 18:06:31.366640 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-544hh" event={"ID":"d2135c03-e50b-4f82-807b-82722514714b","Type":"ContainerStarted","Data":"5b473ad44c9f0f4dc601eb08761a6a8c151d88a35178b887d2591c497fa390c3"} Nov 24 18:06:36 crc kubenswrapper[4777]: I1124 18:06:36.416347 4777 generic.go:334] "Generic (PLEG): container finished" podID="d2135c03-e50b-4f82-807b-82722514714b" containerID="5b473ad44c9f0f4dc601eb08761a6a8c151d88a35178b887d2591c497fa390c3" exitCode=0 Nov 24 18:06:36 crc kubenswrapper[4777]: I1124 18:06:36.416432 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-544hh" event={"ID":"d2135c03-e50b-4f82-807b-82722514714b","Type":"ContainerDied","Data":"5b473ad44c9f0f4dc601eb08761a6a8c151d88a35178b887d2591c497fa390c3"} Nov 24 18:06:37 crc kubenswrapper[4777]: I1124 18:06:37.427242 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-544hh" event={"ID":"d2135c03-e50b-4f82-807b-82722514714b","Type":"ContainerStarted","Data":"55cd58aafa04c9b74df3f4aaadc06f19998b4309c6cb4366aac88566a7447c00"} Nov 24 18:06:37 crc kubenswrapper[4777]: I1124 18:06:37.447754 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-544hh" podStartSLOduration=2.92617447 podStartE2EDuration="10.44773273s" podCreationTimestamp="2025-11-24 18:06:27 +0000 UTC" firstStartedPulling="2025-11-24 18:06:29.341474877 +0000 UTC m=+3947.500669926" lastFinishedPulling="2025-11-24 18:06:36.863033137 +0000 UTC m=+3955.022228186" observedRunningTime="2025-11-24 18:06:37.441935827 +0000 UTC m=+3955.601130886" watchObservedRunningTime="2025-11-24 18:06:37.44773273 +0000 UTC m=+3955.606927779" Nov 24 18:06:37 crc kubenswrapper[4777]: I1124 18:06:37.570877 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-l75dg_0a75da4d-a94d-4639-ad88-0b415b87fe24/cert-manager-controller/0.log" Nov 24 18:06:37 crc kubenswrapper[4777]: I1124 18:06:37.745188 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-596z9_cee47616-4825-45f3-8232-4f46a66c954a/cert-manager-cainjector/0.log" Nov 24 18:06:37 crc kubenswrapper[4777]: I1124 18:06:37.786695 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-9f892_c32a0d11-9b78-4ce6-a4db-d9d07c9f3313/cert-manager-webhook/0.log" Nov 24 18:06:38 crc kubenswrapper[4777]: I1124 18:06:38.311849 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:06:38 crc kubenswrapper[4777]: I1124 18:06:38.311893 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:06:39 crc kubenswrapper[4777]: I1124 18:06:39.355843 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-544hh" podUID="d2135c03-e50b-4f82-807b-82722514714b" containerName="registry-server" probeResult="failure" output=< Nov 24 18:06:39 crc kubenswrapper[4777]: timeout: failed to connect service ":50051" within 1s Nov 24 18:06:39 crc kubenswrapper[4777]: > Nov 24 18:06:39 crc kubenswrapper[4777]: I1124 18:06:39.997997 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:39 crc kubenswrapper[4777]: I1124 18:06:39.998046 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:40 crc kubenswrapper[4777]: I1124 18:06:40.075012 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:40 crc kubenswrapper[4777]: I1124 18:06:40.543751 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:40 crc kubenswrapper[4777]: I1124 18:06:40.619929 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bcmr5"] Nov 24 18:06:42 crc kubenswrapper[4777]: I1124 18:06:42.471908 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bcmr5" podUID="295dc21d-5d67-4ae7-a0b5-676cc0366501" containerName="registry-server" containerID="cri-o://506307466a6cb972e7b9ed3400e0f9a0707ee32f5d99fc7bbf6c2acb68c26c75" gracePeriod=2 Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.193668 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.299981 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wj8q\" (UniqueName: \"kubernetes.io/projected/295dc21d-5d67-4ae7-a0b5-676cc0366501-kube-api-access-2wj8q\") pod \"295dc21d-5d67-4ae7-a0b5-676cc0366501\" (UID: \"295dc21d-5d67-4ae7-a0b5-676cc0366501\") " Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.300434 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295dc21d-5d67-4ae7-a0b5-676cc0366501-catalog-content\") pod \"295dc21d-5d67-4ae7-a0b5-676cc0366501\" (UID: \"295dc21d-5d67-4ae7-a0b5-676cc0366501\") " Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.300487 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295dc21d-5d67-4ae7-a0b5-676cc0366501-utilities\") pod \"295dc21d-5d67-4ae7-a0b5-676cc0366501\" (UID: \"295dc21d-5d67-4ae7-a0b5-676cc0366501\") " Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.301403 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/295dc21d-5d67-4ae7-a0b5-676cc0366501-utilities" (OuterVolumeSpecName: "utilities") pod "295dc21d-5d67-4ae7-a0b5-676cc0366501" (UID: "295dc21d-5d67-4ae7-a0b5-676cc0366501"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.305716 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/295dc21d-5d67-4ae7-a0b5-676cc0366501-kube-api-access-2wj8q" (OuterVolumeSpecName: "kube-api-access-2wj8q") pod "295dc21d-5d67-4ae7-a0b5-676cc0366501" (UID: "295dc21d-5d67-4ae7-a0b5-676cc0366501"). InnerVolumeSpecName "kube-api-access-2wj8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.357409 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/295dc21d-5d67-4ae7-a0b5-676cc0366501-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "295dc21d-5d67-4ae7-a0b5-676cc0366501" (UID: "295dc21d-5d67-4ae7-a0b5-676cc0366501"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.403166 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295dc21d-5d67-4ae7-a0b5-676cc0366501-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.403202 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295dc21d-5d67-4ae7-a0b5-676cc0366501-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.403211 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wj8q\" (UniqueName: \"kubernetes.io/projected/295dc21d-5d67-4ae7-a0b5-676cc0366501-kube-api-access-2wj8q\") on node \"crc\" DevicePath \"\"" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.494002 4777 generic.go:334] "Generic (PLEG): container finished" podID="295dc21d-5d67-4ae7-a0b5-676cc0366501" containerID="506307466a6cb972e7b9ed3400e0f9a0707ee32f5d99fc7bbf6c2acb68c26c75" exitCode=0 Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.494046 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bcmr5" event={"ID":"295dc21d-5d67-4ae7-a0b5-676cc0366501","Type":"ContainerDied","Data":"506307466a6cb972e7b9ed3400e0f9a0707ee32f5d99fc7bbf6c2acb68c26c75"} Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.494071 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bcmr5" event={"ID":"295dc21d-5d67-4ae7-a0b5-676cc0366501","Type":"ContainerDied","Data":"67c56b73f64a921cf480b9392e40754884c11d99fad6f3bcbf0df197aa882722"} Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.494086 4777 scope.go:117] "RemoveContainer" containerID="506307466a6cb972e7b9ed3400e0f9a0707ee32f5d99fc7bbf6c2acb68c26c75" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.494221 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bcmr5" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.533185 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bcmr5"] Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.539689 4777 scope.go:117] "RemoveContainer" containerID="2a50451746d0cdf6b541927860e995d89bfe58c29b2f751594677a7fd23ed797" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.549386 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bcmr5"] Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.571590 4777 scope.go:117] "RemoveContainer" containerID="42235ab3aac7de732755d099a9c472bd4a09ce3be176a72bc11774cf547635f4" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.645001 4777 scope.go:117] "RemoveContainer" containerID="506307466a6cb972e7b9ed3400e0f9a0707ee32f5d99fc7bbf6c2acb68c26c75" Nov 24 18:06:43 crc kubenswrapper[4777]: E1124 18:06:43.645477 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"506307466a6cb972e7b9ed3400e0f9a0707ee32f5d99fc7bbf6c2acb68c26c75\": container with ID starting with 506307466a6cb972e7b9ed3400e0f9a0707ee32f5d99fc7bbf6c2acb68c26c75 not found: ID does not exist" containerID="506307466a6cb972e7b9ed3400e0f9a0707ee32f5d99fc7bbf6c2acb68c26c75" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.645576 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"506307466a6cb972e7b9ed3400e0f9a0707ee32f5d99fc7bbf6c2acb68c26c75"} err="failed to get container status \"506307466a6cb972e7b9ed3400e0f9a0707ee32f5d99fc7bbf6c2acb68c26c75\": rpc error: code = NotFound desc = could not find container \"506307466a6cb972e7b9ed3400e0f9a0707ee32f5d99fc7bbf6c2acb68c26c75\": container with ID starting with 506307466a6cb972e7b9ed3400e0f9a0707ee32f5d99fc7bbf6c2acb68c26c75 not found: ID does not exist" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.645663 4777 scope.go:117] "RemoveContainer" containerID="2a50451746d0cdf6b541927860e995d89bfe58c29b2f751594677a7fd23ed797" Nov 24 18:06:43 crc kubenswrapper[4777]: E1124 18:06:43.645954 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a50451746d0cdf6b541927860e995d89bfe58c29b2f751594677a7fd23ed797\": container with ID starting with 2a50451746d0cdf6b541927860e995d89bfe58c29b2f751594677a7fd23ed797 not found: ID does not exist" containerID="2a50451746d0cdf6b541927860e995d89bfe58c29b2f751594677a7fd23ed797" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.646075 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a50451746d0cdf6b541927860e995d89bfe58c29b2f751594677a7fd23ed797"} err="failed to get container status \"2a50451746d0cdf6b541927860e995d89bfe58c29b2f751594677a7fd23ed797\": rpc error: code = NotFound desc = could not find container \"2a50451746d0cdf6b541927860e995d89bfe58c29b2f751594677a7fd23ed797\": container with ID starting with 2a50451746d0cdf6b541927860e995d89bfe58c29b2f751594677a7fd23ed797 not found: ID does not exist" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.646136 4777 scope.go:117] "RemoveContainer" containerID="42235ab3aac7de732755d099a9c472bd4a09ce3be176a72bc11774cf547635f4" Nov 24 18:06:43 crc kubenswrapper[4777]: E1124 18:06:43.646808 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42235ab3aac7de732755d099a9c472bd4a09ce3be176a72bc11774cf547635f4\": container with ID starting with 42235ab3aac7de732755d099a9c472bd4a09ce3be176a72bc11774cf547635f4 not found: ID does not exist" containerID="42235ab3aac7de732755d099a9c472bd4a09ce3be176a72bc11774cf547635f4" Nov 24 18:06:43 crc kubenswrapper[4777]: I1124 18:06:43.646841 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42235ab3aac7de732755d099a9c472bd4a09ce3be176a72bc11774cf547635f4"} err="failed to get container status \"42235ab3aac7de732755d099a9c472bd4a09ce3be176a72bc11774cf547635f4\": rpc error: code = NotFound desc = could not find container \"42235ab3aac7de732755d099a9c472bd4a09ce3be176a72bc11774cf547635f4\": container with ID starting with 42235ab3aac7de732755d099a9c472bd4a09ce3be176a72bc11774cf547635f4 not found: ID does not exist" Nov 24 18:06:43 crc kubenswrapper[4777]: E1124 18:06:43.771657 4777 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod295dc21d_5d67_4ae7_a0b5_676cc0366501.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod295dc21d_5d67_4ae7_a0b5_676cc0366501.slice/crio-67c56b73f64a921cf480b9392e40754884c11d99fad6f3bcbf0df197aa882722\": RecentStats: unable to find data in memory cache]" Nov 24 18:06:45 crc kubenswrapper[4777]: I1124 18:06:45.258295 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="295dc21d-5d67-4ae7-a0b5-676cc0366501" path="/var/lib/kubelet/pods/295dc21d-5d67-4ae7-a0b5-676cc0366501/volumes" Nov 24 18:06:49 crc kubenswrapper[4777]: I1124 18:06:49.360330 4777 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-544hh" podUID="d2135c03-e50b-4f82-807b-82722514714b" containerName="registry-server" probeResult="failure" output=< Nov 24 18:06:49 crc kubenswrapper[4777]: timeout: failed to connect service ":50051" within 1s Nov 24 18:06:49 crc kubenswrapper[4777]: > Nov 24 18:06:53 crc kubenswrapper[4777]: I1124 18:06:53.555198 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-p6sjq_2f2d18e5-97fa-47f6-9194-760ff54fdddc/nmstate-console-plugin/0.log" Nov 24 18:06:53 crc kubenswrapper[4777]: I1124 18:06:53.633072 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-wnlkc_6aa57fcf-4602-4a14-8f93-d9881aad4c8d/nmstate-handler/0.log" Nov 24 18:06:53 crc kubenswrapper[4777]: I1124 18:06:53.843597 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-m7wgh_2cb2400c-9df2-4e5f-95e9-3e8d22b58d3c/nmstate-metrics/0.log" Nov 24 18:06:53 crc kubenswrapper[4777]: I1124 18:06:53.851516 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-m7wgh_2cb2400c-9df2-4e5f-95e9-3e8d22b58d3c/kube-rbac-proxy/0.log" Nov 24 18:06:54 crc kubenswrapper[4777]: I1124 18:06:54.054601 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-8wtx5_3ffb9c2f-e94c-4fda-8c82-a34c045f9435/nmstate-operator/0.log" Nov 24 18:06:54 crc kubenswrapper[4777]: I1124 18:06:54.143294 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-z59z6_64c0b83e-a764-4494-b384-727c83c914a2/nmstate-webhook/0.log" Nov 24 18:06:58 crc kubenswrapper[4777]: I1124 18:06:58.368429 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:06:58 crc kubenswrapper[4777]: I1124 18:06:58.435584 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:06:59 crc kubenswrapper[4777]: I1124 18:06:59.177928 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-544hh"] Nov 24 18:06:59 crc kubenswrapper[4777]: I1124 18:06:59.672158 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-544hh" podUID="d2135c03-e50b-4f82-807b-82722514714b" containerName="registry-server" containerID="cri-o://55cd58aafa04c9b74df3f4aaadc06f19998b4309c6cb4366aac88566a7447c00" gracePeriod=2 Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.434204 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.543600 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqqvd\" (UniqueName: \"kubernetes.io/projected/d2135c03-e50b-4f82-807b-82722514714b-kube-api-access-fqqvd\") pod \"d2135c03-e50b-4f82-807b-82722514714b\" (UID: \"d2135c03-e50b-4f82-807b-82722514714b\") " Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.543705 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2135c03-e50b-4f82-807b-82722514714b-catalog-content\") pod \"d2135c03-e50b-4f82-807b-82722514714b\" (UID: \"d2135c03-e50b-4f82-807b-82722514714b\") " Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.543808 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2135c03-e50b-4f82-807b-82722514714b-utilities\") pod \"d2135c03-e50b-4f82-807b-82722514714b\" (UID: \"d2135c03-e50b-4f82-807b-82722514714b\") " Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.544720 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2135c03-e50b-4f82-807b-82722514714b-utilities" (OuterVolumeSpecName: "utilities") pod "d2135c03-e50b-4f82-807b-82722514714b" (UID: "d2135c03-e50b-4f82-807b-82722514714b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.549435 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2135c03-e50b-4f82-807b-82722514714b-kube-api-access-fqqvd" (OuterVolumeSpecName: "kube-api-access-fqqvd") pod "d2135c03-e50b-4f82-807b-82722514714b" (UID: "d2135c03-e50b-4f82-807b-82722514714b"). InnerVolumeSpecName "kube-api-access-fqqvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.626835 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2135c03-e50b-4f82-807b-82722514714b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2135c03-e50b-4f82-807b-82722514714b" (UID: "d2135c03-e50b-4f82-807b-82722514714b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.646215 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqqvd\" (UniqueName: \"kubernetes.io/projected/d2135c03-e50b-4f82-807b-82722514714b-kube-api-access-fqqvd\") on node \"crc\" DevicePath \"\"" Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.646248 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2135c03-e50b-4f82-807b-82722514714b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.646258 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2135c03-e50b-4f82-807b-82722514714b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.688066 4777 generic.go:334] "Generic (PLEG): container finished" podID="d2135c03-e50b-4f82-807b-82722514714b" containerID="55cd58aafa04c9b74df3f4aaadc06f19998b4309c6cb4366aac88566a7447c00" exitCode=0 Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.688119 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-544hh" Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.688146 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-544hh" event={"ID":"d2135c03-e50b-4f82-807b-82722514714b","Type":"ContainerDied","Data":"55cd58aafa04c9b74df3f4aaadc06f19998b4309c6cb4366aac88566a7447c00"} Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.688507 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-544hh" event={"ID":"d2135c03-e50b-4f82-807b-82722514714b","Type":"ContainerDied","Data":"3da45c2ea443b2d94753150099821ef97dc570d02d622905358640abf7d80f59"} Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.688529 4777 scope.go:117] "RemoveContainer" containerID="55cd58aafa04c9b74df3f4aaadc06f19998b4309c6cb4366aac88566a7447c00" Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.715636 4777 scope.go:117] "RemoveContainer" containerID="5b473ad44c9f0f4dc601eb08761a6a8c151d88a35178b887d2591c497fa390c3" Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.732509 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-544hh"] Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.739473 4777 scope.go:117] "RemoveContainer" containerID="2cc12217b1faa6c8a5b5f12c83bc3a703f9284a9f5f11da9bb7e06200756e2ea" Nov 24 18:07:00 crc kubenswrapper[4777]: I1124 18:07:00.742182 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-544hh"] Nov 24 18:07:01 crc kubenswrapper[4777]: I1124 18:07:01.262980 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2135c03-e50b-4f82-807b-82722514714b" path="/var/lib/kubelet/pods/d2135c03-e50b-4f82-807b-82722514714b/volumes" Nov 24 18:07:01 crc kubenswrapper[4777]: I1124 18:07:01.264661 4777 scope.go:117] "RemoveContainer" containerID="55cd58aafa04c9b74df3f4aaadc06f19998b4309c6cb4366aac88566a7447c00" Nov 24 18:07:01 crc kubenswrapper[4777]: E1124 18:07:01.265506 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55cd58aafa04c9b74df3f4aaadc06f19998b4309c6cb4366aac88566a7447c00\": container with ID starting with 55cd58aafa04c9b74df3f4aaadc06f19998b4309c6cb4366aac88566a7447c00 not found: ID does not exist" containerID="55cd58aafa04c9b74df3f4aaadc06f19998b4309c6cb4366aac88566a7447c00" Nov 24 18:07:01 crc kubenswrapper[4777]: I1124 18:07:01.265537 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55cd58aafa04c9b74df3f4aaadc06f19998b4309c6cb4366aac88566a7447c00"} err="failed to get container status \"55cd58aafa04c9b74df3f4aaadc06f19998b4309c6cb4366aac88566a7447c00\": rpc error: code = NotFound desc = could not find container \"55cd58aafa04c9b74df3f4aaadc06f19998b4309c6cb4366aac88566a7447c00\": container with ID starting with 55cd58aafa04c9b74df3f4aaadc06f19998b4309c6cb4366aac88566a7447c00 not found: ID does not exist" Nov 24 18:07:01 crc kubenswrapper[4777]: I1124 18:07:01.265564 4777 scope.go:117] "RemoveContainer" containerID="5b473ad44c9f0f4dc601eb08761a6a8c151d88a35178b887d2591c497fa390c3" Nov 24 18:07:01 crc kubenswrapper[4777]: E1124 18:07:01.265933 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b473ad44c9f0f4dc601eb08761a6a8c151d88a35178b887d2591c497fa390c3\": container with ID starting with 5b473ad44c9f0f4dc601eb08761a6a8c151d88a35178b887d2591c497fa390c3 not found: ID does not exist" containerID="5b473ad44c9f0f4dc601eb08761a6a8c151d88a35178b887d2591c497fa390c3" Nov 24 18:07:01 crc kubenswrapper[4777]: I1124 18:07:01.265979 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b473ad44c9f0f4dc601eb08761a6a8c151d88a35178b887d2591c497fa390c3"} err="failed to get container status \"5b473ad44c9f0f4dc601eb08761a6a8c151d88a35178b887d2591c497fa390c3\": rpc error: code = NotFound desc = could not find container \"5b473ad44c9f0f4dc601eb08761a6a8c151d88a35178b887d2591c497fa390c3\": container with ID starting with 5b473ad44c9f0f4dc601eb08761a6a8c151d88a35178b887d2591c497fa390c3 not found: ID does not exist" Nov 24 18:07:01 crc kubenswrapper[4777]: I1124 18:07:01.266000 4777 scope.go:117] "RemoveContainer" containerID="2cc12217b1faa6c8a5b5f12c83bc3a703f9284a9f5f11da9bb7e06200756e2ea" Nov 24 18:07:01 crc kubenswrapper[4777]: E1124 18:07:01.266760 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cc12217b1faa6c8a5b5f12c83bc3a703f9284a9f5f11da9bb7e06200756e2ea\": container with ID starting with 2cc12217b1faa6c8a5b5f12c83bc3a703f9284a9f5f11da9bb7e06200756e2ea not found: ID does not exist" containerID="2cc12217b1faa6c8a5b5f12c83bc3a703f9284a9f5f11da9bb7e06200756e2ea" Nov 24 18:07:01 crc kubenswrapper[4777]: I1124 18:07:01.266790 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cc12217b1faa6c8a5b5f12c83bc3a703f9284a9f5f11da9bb7e06200756e2ea"} err="failed to get container status \"2cc12217b1faa6c8a5b5f12c83bc3a703f9284a9f5f11da9bb7e06200756e2ea\": rpc error: code = NotFound desc = could not find container \"2cc12217b1faa6c8a5b5f12c83bc3a703f9284a9f5f11da9bb7e06200756e2ea\": container with ID starting with 2cc12217b1faa6c8a5b5f12c83bc3a703f9284a9f5f11da9bb7e06200756e2ea not found: ID does not exist" Nov 24 18:07:10 crc kubenswrapper[4777]: I1124 18:07:10.570620 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6ffccfff7d-t8ltx_d1c8f636-6ede-4469-b270-63d3823d6507/kube-rbac-proxy/0.log" Nov 24 18:07:10 crc kubenswrapper[4777]: I1124 18:07:10.678064 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6ffccfff7d-t8ltx_d1c8f636-6ede-4469-b270-63d3823d6507/manager/0.log" Nov 24 18:07:25 crc kubenswrapper[4777]: I1124 18:07:25.237854 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-z9z24_b4a78161-cf5f-4307-9adb-a05de51d620f/kube-rbac-proxy/0.log" Nov 24 18:07:25 crc kubenswrapper[4777]: I1124 18:07:25.289458 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-z9z24_b4a78161-cf5f-4307-9adb-a05de51d620f/controller/0.log" Nov 24 18:07:25 crc kubenswrapper[4777]: I1124 18:07:25.410659 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-frr-files/0.log" Nov 24 18:07:25 crc kubenswrapper[4777]: I1124 18:07:25.619587 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-frr-files/0.log" Nov 24 18:07:25 crc kubenswrapper[4777]: I1124 18:07:25.695376 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-reloader/0.log" Nov 24 18:07:25 crc kubenswrapper[4777]: I1124 18:07:25.713079 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-reloader/0.log" Nov 24 18:07:25 crc kubenswrapper[4777]: I1124 18:07:25.713193 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-metrics/0.log" Nov 24 18:07:25 crc kubenswrapper[4777]: I1124 18:07:25.890330 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-metrics/0.log" Nov 24 18:07:25 crc kubenswrapper[4777]: I1124 18:07:25.897362 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-frr-files/0.log" Nov 24 18:07:25 crc kubenswrapper[4777]: I1124 18:07:25.903531 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-reloader/0.log" Nov 24 18:07:25 crc kubenswrapper[4777]: I1124 18:07:25.915572 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-metrics/0.log" Nov 24 18:07:26 crc kubenswrapper[4777]: I1124 18:07:26.114455 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-frr-files/0.log" Nov 24 18:07:26 crc kubenswrapper[4777]: I1124 18:07:26.117774 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/controller/0.log" Nov 24 18:07:26 crc kubenswrapper[4777]: I1124 18:07:26.127639 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-reloader/0.log" Nov 24 18:07:26 crc kubenswrapper[4777]: I1124 18:07:26.158909 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/cp-metrics/0.log" Nov 24 18:07:26 crc kubenswrapper[4777]: I1124 18:07:26.309279 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/frr-metrics/0.log" Nov 24 18:07:26 crc kubenswrapper[4777]: I1124 18:07:26.362218 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/kube-rbac-proxy-frr/0.log" Nov 24 18:07:26 crc kubenswrapper[4777]: I1124 18:07:26.427810 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/kube-rbac-proxy/0.log" Nov 24 18:07:26 crc kubenswrapper[4777]: I1124 18:07:26.646352 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/reloader/0.log" Nov 24 18:07:26 crc kubenswrapper[4777]: I1124 18:07:26.703215 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-86vqd_2e13d83f-3c71-4981-9cbb-9d0ef94c1a5a/frr-k8s-webhook-server/0.log" Nov 24 18:07:26 crc kubenswrapper[4777]: I1124 18:07:26.957345 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7444c76584-h5nzb_364328a8-4ddb-4c3c-96bd-3a87ef271678/manager/0.log" Nov 24 18:07:27 crc kubenswrapper[4777]: I1124 18:07:27.108946 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7578594d7b-kn9dd_61277f6c-5e55-4fdb-a482-e9b3a91edd7f/webhook-server/0.log" Nov 24 18:07:27 crc kubenswrapper[4777]: I1124 18:07:27.234078 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-pz9ts_5ae956fa-5a45-4248-b300-d802aa171d52/kube-rbac-proxy/0.log" Nov 24 18:07:27 crc kubenswrapper[4777]: I1124 18:07:27.603395 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c95bf_0dd8e42b-e649-438d-bd92-e35a3ed06d4a/frr/0.log" Nov 24 18:07:27 crc kubenswrapper[4777]: I1124 18:07:27.707138 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-pz9ts_5ae956fa-5a45-4248-b300-d802aa171d52/speaker/0.log" Nov 24 18:07:42 crc kubenswrapper[4777]: I1124 18:07:42.430350 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c_1ffbfe00-f54c-446d-bc90-d18f24f51f82/util/0.log" Nov 24 18:07:42 crc kubenswrapper[4777]: I1124 18:07:42.691168 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c_1ffbfe00-f54c-446d-bc90-d18f24f51f82/pull/0.log" Nov 24 18:07:42 crc kubenswrapper[4777]: I1124 18:07:42.695486 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c_1ffbfe00-f54c-446d-bc90-d18f24f51f82/util/0.log" Nov 24 18:07:42 crc kubenswrapper[4777]: I1124 18:07:42.726730 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c_1ffbfe00-f54c-446d-bc90-d18f24f51f82/pull/0.log" Nov 24 18:07:42 crc kubenswrapper[4777]: I1124 18:07:42.875501 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c_1ffbfe00-f54c-446d-bc90-d18f24f51f82/util/0.log" Nov 24 18:07:42 crc kubenswrapper[4777]: I1124 18:07:42.918095 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c_1ffbfe00-f54c-446d-bc90-d18f24f51f82/pull/0.log" Nov 24 18:07:42 crc kubenswrapper[4777]: I1124 18:07:42.957043 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694hg65c_1ffbfe00-f54c-446d-bc90-d18f24f51f82/extract/0.log" Nov 24 18:07:43 crc kubenswrapper[4777]: I1124 18:07:43.113657 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd_edece1c5-1d2f-4a61-9591-7313cd41d58d/util/0.log" Nov 24 18:07:43 crc kubenswrapper[4777]: I1124 18:07:43.280308 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd_edece1c5-1d2f-4a61-9591-7313cd41d58d/util/0.log" Nov 24 18:07:43 crc kubenswrapper[4777]: I1124 18:07:43.289020 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd_edece1c5-1d2f-4a61-9591-7313cd41d58d/pull/0.log" Nov 24 18:07:43 crc kubenswrapper[4777]: I1124 18:07:43.309101 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd_edece1c5-1d2f-4a61-9591-7313cd41d58d/pull/0.log" Nov 24 18:07:43 crc kubenswrapper[4777]: I1124 18:07:43.449790 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd_edece1c5-1d2f-4a61-9591-7313cd41d58d/util/0.log" Nov 24 18:07:43 crc kubenswrapper[4777]: I1124 18:07:43.475013 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd_edece1c5-1d2f-4a61-9591-7313cd41d58d/extract/0.log" Nov 24 18:07:43 crc kubenswrapper[4777]: I1124 18:07:43.476871 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303fz5fd_edece1c5-1d2f-4a61-9591-7313cd41d58d/pull/0.log" Nov 24 18:07:43 crc kubenswrapper[4777]: I1124 18:07:43.964216 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww_213b5cf0-7684-4331-b5be-28fd188e0db3/util/0.log" Nov 24 18:07:44 crc kubenswrapper[4777]: I1124 18:07:44.127381 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww_213b5cf0-7684-4331-b5be-28fd188e0db3/pull/0.log" Nov 24 18:07:44 crc kubenswrapper[4777]: I1124 18:07:44.182239 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww_213b5cf0-7684-4331-b5be-28fd188e0db3/pull/0.log" Nov 24 18:07:44 crc kubenswrapper[4777]: I1124 18:07:44.216182 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww_213b5cf0-7684-4331-b5be-28fd188e0db3/util/0.log" Nov 24 18:07:44 crc kubenswrapper[4777]: I1124 18:07:44.375072 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww_213b5cf0-7684-4331-b5be-28fd188e0db3/util/0.log" Nov 24 18:07:44 crc kubenswrapper[4777]: I1124 18:07:44.376694 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww_213b5cf0-7684-4331-b5be-28fd188e0db3/pull/0.log" Nov 24 18:07:44 crc kubenswrapper[4777]: I1124 18:07:44.396663 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edbmww_213b5cf0-7684-4331-b5be-28fd188e0db3/extract/0.log" Nov 24 18:07:44 crc kubenswrapper[4777]: I1124 18:07:44.537542 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9_9194629c-e5c7-4ef7-8b08-67ae8d58a1a2/util/0.log" Nov 24 18:07:44 crc kubenswrapper[4777]: I1124 18:07:44.744111 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9_9194629c-e5c7-4ef7-8b08-67ae8d58a1a2/util/0.log" Nov 24 18:07:44 crc kubenswrapper[4777]: I1124 18:07:44.754054 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9_9194629c-e5c7-4ef7-8b08-67ae8d58a1a2/pull/0.log" Nov 24 18:07:44 crc kubenswrapper[4777]: I1124 18:07:44.773735 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9_9194629c-e5c7-4ef7-8b08-67ae8d58a1a2/pull/0.log" Nov 24 18:07:44 crc kubenswrapper[4777]: I1124 18:07:44.998342 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9_9194629c-e5c7-4ef7-8b08-67ae8d58a1a2/extract/0.log" Nov 24 18:07:45 crc kubenswrapper[4777]: I1124 18:07:45.070304 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9_9194629c-e5c7-4ef7-8b08-67ae8d58a1a2/pull/0.log" Nov 24 18:07:45 crc kubenswrapper[4777]: I1124 18:07:45.082214 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92102wnx9_9194629c-e5c7-4ef7-8b08-67ae8d58a1a2/util/0.log" Nov 24 18:07:45 crc kubenswrapper[4777]: I1124 18:07:45.210566 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5m4fr_6aa2fa6f-bd34-4154-b666-455e7fea71ae/extract-utilities/0.log" Nov 24 18:07:45 crc kubenswrapper[4777]: I1124 18:07:45.425218 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5m4fr_6aa2fa6f-bd34-4154-b666-455e7fea71ae/extract-content/0.log" Nov 24 18:07:45 crc kubenswrapper[4777]: I1124 18:07:45.458258 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5m4fr_6aa2fa6f-bd34-4154-b666-455e7fea71ae/extract-content/0.log" Nov 24 18:07:45 crc kubenswrapper[4777]: I1124 18:07:45.458546 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5m4fr_6aa2fa6f-bd34-4154-b666-455e7fea71ae/extract-utilities/0.log" Nov 24 18:07:45 crc kubenswrapper[4777]: I1124 18:07:45.590665 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5m4fr_6aa2fa6f-bd34-4154-b666-455e7fea71ae/extract-utilities/0.log" Nov 24 18:07:45 crc kubenswrapper[4777]: I1124 18:07:45.657095 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5m4fr_6aa2fa6f-bd34-4154-b666-455e7fea71ae/extract-content/0.log" Nov 24 18:07:45 crc kubenswrapper[4777]: I1124 18:07:45.767186 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h54gs_30623966-678a-4124-bc99-02d4ccfacb85/extract-utilities/0.log" Nov 24 18:07:46 crc kubenswrapper[4777]: I1124 18:07:46.126257 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h54gs_30623966-678a-4124-bc99-02d4ccfacb85/extract-utilities/0.log" Nov 24 18:07:46 crc kubenswrapper[4777]: I1124 18:07:46.142281 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h54gs_30623966-678a-4124-bc99-02d4ccfacb85/extract-content/0.log" Nov 24 18:07:46 crc kubenswrapper[4777]: I1124 18:07:46.187075 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h54gs_30623966-678a-4124-bc99-02d4ccfacb85/extract-content/0.log" Nov 24 18:07:46 crc kubenswrapper[4777]: I1124 18:07:46.248449 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5m4fr_6aa2fa6f-bd34-4154-b666-455e7fea71ae/registry-server/0.log" Nov 24 18:07:46 crc kubenswrapper[4777]: I1124 18:07:46.430556 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h54gs_30623966-678a-4124-bc99-02d4ccfacb85/extract-utilities/0.log" Nov 24 18:07:46 crc kubenswrapper[4777]: I1124 18:07:46.458626 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h54gs_30623966-678a-4124-bc99-02d4ccfacb85/extract-content/0.log" Nov 24 18:07:46 crc kubenswrapper[4777]: I1124 18:07:46.527855 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4_539a904e-3754-42f1-95ec-13140964520d/util/0.log" Nov 24 18:07:46 crc kubenswrapper[4777]: I1124 18:07:46.666611 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4_539a904e-3754-42f1-95ec-13140964520d/util/0.log" Nov 24 18:07:46 crc kubenswrapper[4777]: I1124 18:07:46.682385 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4_539a904e-3754-42f1-95ec-13140964520d/pull/0.log" Nov 24 18:07:46 crc kubenswrapper[4777]: I1124 18:07:46.689595 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4_539a904e-3754-42f1-95ec-13140964520d/pull/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.018906 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4_539a904e-3754-42f1-95ec-13140964520d/pull/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.056108 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4_539a904e-3754-42f1-95ec-13140964520d/extract/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.060604 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h54gs_30623966-678a-4124-bc99-02d4ccfacb85/registry-server/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.074528 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6cc5f4_539a904e-3754-42f1-95ec-13140964520d/util/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.236262 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-5dsxp_3dea83d3-9f3d-4f0d-ad21-ffb6a0f997da/marketplace-operator/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.246222 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hwftb_27733c5c-8536-4f12-8aa0-e5a9d16e8f88/extract-utilities/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.486799 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hwftb_27733c5c-8536-4f12-8aa0-e5a9d16e8f88/extract-utilities/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.497011 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hwftb_27733c5c-8536-4f12-8aa0-e5a9d16e8f88/extract-content/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.512105 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hwftb_27733c5c-8536-4f12-8aa0-e5a9d16e8f88/extract-content/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.656947 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hwftb_27733c5c-8536-4f12-8aa0-e5a9d16e8f88/extract-content/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.707995 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hwftb_27733c5c-8536-4f12-8aa0-e5a9d16e8f88/extract-utilities/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.757614 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g7f5l_1af51c54-3a40-406e-8743-32b0c04d018f/extract-utilities/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.829272 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hwftb_27733c5c-8536-4f12-8aa0-e5a9d16e8f88/registry-server/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.940192 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g7f5l_1af51c54-3a40-406e-8743-32b0c04d018f/extract-content/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.941354 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g7f5l_1af51c54-3a40-406e-8743-32b0c04d018f/extract-utilities/0.log" Nov 24 18:07:47 crc kubenswrapper[4777]: I1124 18:07:47.983096 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g7f5l_1af51c54-3a40-406e-8743-32b0c04d018f/extract-content/0.log" Nov 24 18:07:48 crc kubenswrapper[4777]: I1124 18:07:48.151763 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g7f5l_1af51c54-3a40-406e-8743-32b0c04d018f/extract-content/0.log" Nov 24 18:07:48 crc kubenswrapper[4777]: I1124 18:07:48.165180 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g7f5l_1af51c54-3a40-406e-8743-32b0c04d018f/extract-utilities/0.log" Nov 24 18:07:48 crc kubenswrapper[4777]: I1124 18:07:48.808727 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g7f5l_1af51c54-3a40-406e-8743-32b0c04d018f/registry-server/0.log" Nov 24 18:08:02 crc kubenswrapper[4777]: I1124 18:08:02.458470 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-wsbfx_b08a2563-7e34-40f7-b9b0-052824422cff/prometheus-operator/0.log" Nov 24 18:08:03 crc kubenswrapper[4777]: I1124 18:08:03.295787 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-58cb946cb6-9nb4r_dd6addca-3fb6-4204-ae4e-2ed8d1326699/prometheus-operator-admission-webhook/0.log" Nov 24 18:08:03 crc kubenswrapper[4777]: I1124 18:08:03.302758 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-58cb946cb6-jhzjq_67418bcd-f0ba-4526-afc1-8420cb642ccf/prometheus-operator-admission-webhook/0.log" Nov 24 18:08:03 crc kubenswrapper[4777]: I1124 18:08:03.473550 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-hdxt8_986d062f-3210-471c-a2db-07e080b5d449/operator/0.log" Nov 24 18:08:03 crc kubenswrapper[4777]: I1124 18:08:03.573789 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-22cbh_c0fe98e7-9ad1-426e-a17a-ad471a7f73af/perses-operator/0.log" Nov 24 18:08:14 crc kubenswrapper[4777]: I1124 18:08:14.269923 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:08:14 crc kubenswrapper[4777]: I1124 18:08:14.270397 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:08:16 crc kubenswrapper[4777]: I1124 18:08:16.986140 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6ffccfff7d-t8ltx_d1c8f636-6ede-4469-b270-63d3823d6507/kube-rbac-proxy/0.log" Nov 24 18:08:17 crc kubenswrapper[4777]: I1124 18:08:17.012129 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6ffccfff7d-t8ltx_d1c8f636-6ede-4469-b270-63d3823d6507/manager/0.log" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.319245 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-22hvr"] Nov 24 18:08:32 crc kubenswrapper[4777]: E1124 18:08:32.320173 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295dc21d-5d67-4ae7-a0b5-676cc0366501" containerName="extract-content" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.320186 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="295dc21d-5d67-4ae7-a0b5-676cc0366501" containerName="extract-content" Nov 24 18:08:32 crc kubenswrapper[4777]: E1124 18:08:32.320199 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2135c03-e50b-4f82-807b-82722514714b" containerName="registry-server" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.320205 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2135c03-e50b-4f82-807b-82722514714b" containerName="registry-server" Nov 24 18:08:32 crc kubenswrapper[4777]: E1124 18:08:32.320216 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295dc21d-5d67-4ae7-a0b5-676cc0366501" containerName="registry-server" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.320224 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="295dc21d-5d67-4ae7-a0b5-676cc0366501" containerName="registry-server" Nov 24 18:08:32 crc kubenswrapper[4777]: E1124 18:08:32.320256 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295dc21d-5d67-4ae7-a0b5-676cc0366501" containerName="extract-utilities" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.320262 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="295dc21d-5d67-4ae7-a0b5-676cc0366501" containerName="extract-utilities" Nov 24 18:08:32 crc kubenswrapper[4777]: E1124 18:08:32.320277 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2135c03-e50b-4f82-807b-82722514714b" containerName="extract-content" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.320282 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2135c03-e50b-4f82-807b-82722514714b" containerName="extract-content" Nov 24 18:08:32 crc kubenswrapper[4777]: E1124 18:08:32.320317 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2135c03-e50b-4f82-807b-82722514714b" containerName="extract-utilities" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.320322 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2135c03-e50b-4f82-807b-82722514714b" containerName="extract-utilities" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.320534 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2135c03-e50b-4f82-807b-82722514714b" containerName="registry-server" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.320548 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="295dc21d-5d67-4ae7-a0b5-676cc0366501" containerName="registry-server" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.322065 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.334947 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-22hvr"] Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.416149 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7e48980-5467-4d69-9a43-3437ca6aa0b7-catalog-content\") pod \"community-operators-22hvr\" (UID: \"e7e48980-5467-4d69-9a43-3437ca6aa0b7\") " pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.416228 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwch5\" (UniqueName: \"kubernetes.io/projected/e7e48980-5467-4d69-9a43-3437ca6aa0b7-kube-api-access-lwch5\") pod \"community-operators-22hvr\" (UID: \"e7e48980-5467-4d69-9a43-3437ca6aa0b7\") " pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.416394 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7e48980-5467-4d69-9a43-3437ca6aa0b7-utilities\") pod \"community-operators-22hvr\" (UID: \"e7e48980-5467-4d69-9a43-3437ca6aa0b7\") " pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.518041 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwch5\" (UniqueName: \"kubernetes.io/projected/e7e48980-5467-4d69-9a43-3437ca6aa0b7-kube-api-access-lwch5\") pod \"community-operators-22hvr\" (UID: \"e7e48980-5467-4d69-9a43-3437ca6aa0b7\") " pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.518206 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7e48980-5467-4d69-9a43-3437ca6aa0b7-utilities\") pod \"community-operators-22hvr\" (UID: \"e7e48980-5467-4d69-9a43-3437ca6aa0b7\") " pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.518274 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7e48980-5467-4d69-9a43-3437ca6aa0b7-catalog-content\") pod \"community-operators-22hvr\" (UID: \"e7e48980-5467-4d69-9a43-3437ca6aa0b7\") " pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.518712 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7e48980-5467-4d69-9a43-3437ca6aa0b7-catalog-content\") pod \"community-operators-22hvr\" (UID: \"e7e48980-5467-4d69-9a43-3437ca6aa0b7\") " pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.518923 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7e48980-5467-4d69-9a43-3437ca6aa0b7-utilities\") pod \"community-operators-22hvr\" (UID: \"e7e48980-5467-4d69-9a43-3437ca6aa0b7\") " pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.850460 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwch5\" (UniqueName: \"kubernetes.io/projected/e7e48980-5467-4d69-9a43-3437ca6aa0b7-kube-api-access-lwch5\") pod \"community-operators-22hvr\" (UID: \"e7e48980-5467-4d69-9a43-3437ca6aa0b7\") " pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:32 crc kubenswrapper[4777]: I1124 18:08:32.958178 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:33 crc kubenswrapper[4777]: I1124 18:08:33.564299 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-22hvr"] Nov 24 18:08:33 crc kubenswrapper[4777]: I1124 18:08:33.601406 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-22hvr" event={"ID":"e7e48980-5467-4d69-9a43-3437ca6aa0b7","Type":"ContainerStarted","Data":"1bdfdada4b5748dcb71f1d1e83bc21cc3a2f2e14fe10cac0e7eb9343ba93d137"} Nov 24 18:08:33 crc kubenswrapper[4777]: E1124 18:08:33.880263 4777 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.9:37286->38.102.83.9:37723: write tcp 38.102.83.9:37286->38.102.83.9:37723: write: broken pipe Nov 24 18:08:34 crc kubenswrapper[4777]: I1124 18:08:34.612076 4777 generic.go:334] "Generic (PLEG): container finished" podID="e7e48980-5467-4d69-9a43-3437ca6aa0b7" containerID="58652ba1647ce9f7730d0b655a595a3f08e6fd835cdc7c258d229a08f788ba8c" exitCode=0 Nov 24 18:08:34 crc kubenswrapper[4777]: I1124 18:08:34.612126 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-22hvr" event={"ID":"e7e48980-5467-4d69-9a43-3437ca6aa0b7","Type":"ContainerDied","Data":"58652ba1647ce9f7730d0b655a595a3f08e6fd835cdc7c258d229a08f788ba8c"} Nov 24 18:08:35 crc kubenswrapper[4777]: I1124 18:08:35.625532 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-22hvr" event={"ID":"e7e48980-5467-4d69-9a43-3437ca6aa0b7","Type":"ContainerStarted","Data":"dff35816c7601dc95a4edad16cb75807986b0d220328b67ddbd6c9206a19243c"} Nov 24 18:08:37 crc kubenswrapper[4777]: I1124 18:08:37.643921 4777 generic.go:334] "Generic (PLEG): container finished" podID="e7e48980-5467-4d69-9a43-3437ca6aa0b7" containerID="dff35816c7601dc95a4edad16cb75807986b0d220328b67ddbd6c9206a19243c" exitCode=0 Nov 24 18:08:37 crc kubenswrapper[4777]: I1124 18:08:37.644030 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-22hvr" event={"ID":"e7e48980-5467-4d69-9a43-3437ca6aa0b7","Type":"ContainerDied","Data":"dff35816c7601dc95a4edad16cb75807986b0d220328b67ddbd6c9206a19243c"} Nov 24 18:08:38 crc kubenswrapper[4777]: I1124 18:08:38.659490 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-22hvr" event={"ID":"e7e48980-5467-4d69-9a43-3437ca6aa0b7","Type":"ContainerStarted","Data":"c0441b2d5f96a55fb05a5946a895a564a13aa407ab5296b1ec2e6a60e81ce02e"} Nov 24 18:08:38 crc kubenswrapper[4777]: I1124 18:08:38.682375 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-22hvr" podStartSLOduration=2.95185281 podStartE2EDuration="6.682352683s" podCreationTimestamp="2025-11-24 18:08:32 +0000 UTC" firstStartedPulling="2025-11-24 18:08:34.614262964 +0000 UTC m=+4072.773458013" lastFinishedPulling="2025-11-24 18:08:38.344762837 +0000 UTC m=+4076.503957886" observedRunningTime="2025-11-24 18:08:38.680477141 +0000 UTC m=+4076.839672200" watchObservedRunningTime="2025-11-24 18:08:38.682352683 +0000 UTC m=+4076.841547742" Nov 24 18:08:42 crc kubenswrapper[4777]: I1124 18:08:42.958389 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:42 crc kubenswrapper[4777]: I1124 18:08:42.959113 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:43 crc kubenswrapper[4777]: I1124 18:08:43.099857 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:43 crc kubenswrapper[4777]: I1124 18:08:43.799348 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:43 crc kubenswrapper[4777]: I1124 18:08:43.850917 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-22hvr"] Nov 24 18:08:44 crc kubenswrapper[4777]: I1124 18:08:44.270222 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:08:44 crc kubenswrapper[4777]: I1124 18:08:44.270291 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:08:45 crc kubenswrapper[4777]: I1124 18:08:45.747317 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-22hvr" podUID="e7e48980-5467-4d69-9a43-3437ca6aa0b7" containerName="registry-server" containerID="cri-o://c0441b2d5f96a55fb05a5946a895a564a13aa407ab5296b1ec2e6a60e81ce02e" gracePeriod=2 Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.498828 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.619657 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwch5\" (UniqueName: \"kubernetes.io/projected/e7e48980-5467-4d69-9a43-3437ca6aa0b7-kube-api-access-lwch5\") pod \"e7e48980-5467-4d69-9a43-3437ca6aa0b7\" (UID: \"e7e48980-5467-4d69-9a43-3437ca6aa0b7\") " Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.619717 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7e48980-5467-4d69-9a43-3437ca6aa0b7-utilities\") pod \"e7e48980-5467-4d69-9a43-3437ca6aa0b7\" (UID: \"e7e48980-5467-4d69-9a43-3437ca6aa0b7\") " Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.620101 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7e48980-5467-4d69-9a43-3437ca6aa0b7-catalog-content\") pod \"e7e48980-5467-4d69-9a43-3437ca6aa0b7\" (UID: \"e7e48980-5467-4d69-9a43-3437ca6aa0b7\") " Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.621026 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7e48980-5467-4d69-9a43-3437ca6aa0b7-utilities" (OuterVolumeSpecName: "utilities") pod "e7e48980-5467-4d69-9a43-3437ca6aa0b7" (UID: "e7e48980-5467-4d69-9a43-3437ca6aa0b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.626754 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e48980-5467-4d69-9a43-3437ca6aa0b7-kube-api-access-lwch5" (OuterVolumeSpecName: "kube-api-access-lwch5") pod "e7e48980-5467-4d69-9a43-3437ca6aa0b7" (UID: "e7e48980-5467-4d69-9a43-3437ca6aa0b7"). InnerVolumeSpecName "kube-api-access-lwch5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.724831 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwch5\" (UniqueName: \"kubernetes.io/projected/e7e48980-5467-4d69-9a43-3437ca6aa0b7-kube-api-access-lwch5\") on node \"crc\" DevicePath \"\"" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.728411 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7e48980-5467-4d69-9a43-3437ca6aa0b7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.770821 4777 generic.go:334] "Generic (PLEG): container finished" podID="e7e48980-5467-4d69-9a43-3437ca6aa0b7" containerID="c0441b2d5f96a55fb05a5946a895a564a13aa407ab5296b1ec2e6a60e81ce02e" exitCode=0 Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.770875 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-22hvr" event={"ID":"e7e48980-5467-4d69-9a43-3437ca6aa0b7","Type":"ContainerDied","Data":"c0441b2d5f96a55fb05a5946a895a564a13aa407ab5296b1ec2e6a60e81ce02e"} Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.770903 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-22hvr" event={"ID":"e7e48980-5467-4d69-9a43-3437ca6aa0b7","Type":"ContainerDied","Data":"1bdfdada4b5748dcb71f1d1e83bc21cc3a2f2e14fe10cac0e7eb9343ba93d137"} Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.770922 4777 scope.go:117] "RemoveContainer" containerID="c0441b2d5f96a55fb05a5946a895a564a13aa407ab5296b1ec2e6a60e81ce02e" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.771106 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-22hvr" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.794190 4777 scope.go:117] "RemoveContainer" containerID="dff35816c7601dc95a4edad16cb75807986b0d220328b67ddbd6c9206a19243c" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.822592 4777 scope.go:117] "RemoveContainer" containerID="58652ba1647ce9f7730d0b655a595a3f08e6fd835cdc7c258d229a08f788ba8c" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.914308 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7e48980-5467-4d69-9a43-3437ca6aa0b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7e48980-5467-4d69-9a43-3437ca6aa0b7" (UID: "e7e48980-5467-4d69-9a43-3437ca6aa0b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.937192 4777 scope.go:117] "RemoveContainer" containerID="c0441b2d5f96a55fb05a5946a895a564a13aa407ab5296b1ec2e6a60e81ce02e" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.938610 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7e48980-5467-4d69-9a43-3437ca6aa0b7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:08:46 crc kubenswrapper[4777]: E1124 18:08:46.938841 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0441b2d5f96a55fb05a5946a895a564a13aa407ab5296b1ec2e6a60e81ce02e\": container with ID starting with c0441b2d5f96a55fb05a5946a895a564a13aa407ab5296b1ec2e6a60e81ce02e not found: ID does not exist" containerID="c0441b2d5f96a55fb05a5946a895a564a13aa407ab5296b1ec2e6a60e81ce02e" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.938886 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0441b2d5f96a55fb05a5946a895a564a13aa407ab5296b1ec2e6a60e81ce02e"} err="failed to get container status \"c0441b2d5f96a55fb05a5946a895a564a13aa407ab5296b1ec2e6a60e81ce02e\": rpc error: code = NotFound desc = could not find container \"c0441b2d5f96a55fb05a5946a895a564a13aa407ab5296b1ec2e6a60e81ce02e\": container with ID starting with c0441b2d5f96a55fb05a5946a895a564a13aa407ab5296b1ec2e6a60e81ce02e not found: ID does not exist" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.938915 4777 scope.go:117] "RemoveContainer" containerID="dff35816c7601dc95a4edad16cb75807986b0d220328b67ddbd6c9206a19243c" Nov 24 18:08:46 crc kubenswrapper[4777]: E1124 18:08:46.939252 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dff35816c7601dc95a4edad16cb75807986b0d220328b67ddbd6c9206a19243c\": container with ID starting with dff35816c7601dc95a4edad16cb75807986b0d220328b67ddbd6c9206a19243c not found: ID does not exist" containerID="dff35816c7601dc95a4edad16cb75807986b0d220328b67ddbd6c9206a19243c" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.939281 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dff35816c7601dc95a4edad16cb75807986b0d220328b67ddbd6c9206a19243c"} err="failed to get container status \"dff35816c7601dc95a4edad16cb75807986b0d220328b67ddbd6c9206a19243c\": rpc error: code = NotFound desc = could not find container \"dff35816c7601dc95a4edad16cb75807986b0d220328b67ddbd6c9206a19243c\": container with ID starting with dff35816c7601dc95a4edad16cb75807986b0d220328b67ddbd6c9206a19243c not found: ID does not exist" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.939297 4777 scope.go:117] "RemoveContainer" containerID="58652ba1647ce9f7730d0b655a595a3f08e6fd835cdc7c258d229a08f788ba8c" Nov 24 18:08:46 crc kubenswrapper[4777]: E1124 18:08:46.939546 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58652ba1647ce9f7730d0b655a595a3f08e6fd835cdc7c258d229a08f788ba8c\": container with ID starting with 58652ba1647ce9f7730d0b655a595a3f08e6fd835cdc7c258d229a08f788ba8c not found: ID does not exist" containerID="58652ba1647ce9f7730d0b655a595a3f08e6fd835cdc7c258d229a08f788ba8c" Nov 24 18:08:46 crc kubenswrapper[4777]: I1124 18:08:46.939591 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58652ba1647ce9f7730d0b655a595a3f08e6fd835cdc7c258d229a08f788ba8c"} err="failed to get container status \"58652ba1647ce9f7730d0b655a595a3f08e6fd835cdc7c258d229a08f788ba8c\": rpc error: code = NotFound desc = could not find container \"58652ba1647ce9f7730d0b655a595a3f08e6fd835cdc7c258d229a08f788ba8c\": container with ID starting with 58652ba1647ce9f7730d0b655a595a3f08e6fd835cdc7c258d229a08f788ba8c not found: ID does not exist" Nov 24 18:08:47 crc kubenswrapper[4777]: I1124 18:08:47.112048 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-22hvr"] Nov 24 18:08:47 crc kubenswrapper[4777]: I1124 18:08:47.129782 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-22hvr"] Nov 24 18:08:47 crc kubenswrapper[4777]: E1124 18:08:47.229501 4777 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7e48980_5467_4d69_9a43_3437ca6aa0b7.slice/crio-1bdfdada4b5748dcb71f1d1e83bc21cc3a2f2e14fe10cac0e7eb9343ba93d137\": RecentStats: unable to find data in memory cache]" Nov 24 18:08:47 crc kubenswrapper[4777]: I1124 18:08:47.259213 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e48980-5467-4d69-9a43-3437ca6aa0b7" path="/var/lib/kubelet/pods/e7e48980-5467-4d69-9a43-3437ca6aa0b7/volumes" Nov 24 18:09:14 crc kubenswrapper[4777]: I1124 18:09:14.270275 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:09:14 crc kubenswrapper[4777]: I1124 18:09:14.270880 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:09:14 crc kubenswrapper[4777]: I1124 18:09:14.270935 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 18:09:14 crc kubenswrapper[4777]: I1124 18:09:14.271816 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1b72042dc72b04f0b4e64de35ea4d6163e72725033a38990bed0f8f82baa07f0"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 18:09:14 crc kubenswrapper[4777]: I1124 18:09:14.271885 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://1b72042dc72b04f0b4e64de35ea4d6163e72725033a38990bed0f8f82baa07f0" gracePeriod=600 Nov 24 18:09:15 crc kubenswrapper[4777]: I1124 18:09:15.033000 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="1b72042dc72b04f0b4e64de35ea4d6163e72725033a38990bed0f8f82baa07f0" exitCode=0 Nov 24 18:09:15 crc kubenswrapper[4777]: I1124 18:09:15.033051 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"1b72042dc72b04f0b4e64de35ea4d6163e72725033a38990bed0f8f82baa07f0"} Nov 24 18:09:15 crc kubenswrapper[4777]: I1124 18:09:15.033088 4777 scope.go:117] "RemoveContainer" containerID="20ea51989d308cca5ee0222497caabfcc3f5073fea72d448cbc7bf1900600dd6" Nov 24 18:09:16 crc kubenswrapper[4777]: I1124 18:09:16.045068 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerStarted","Data":"8e73f0ac9e582e0fd5e9016a5582a523e4e3ccdb50f6f1d6e5751489d53ee9df"} Nov 24 18:10:01 crc kubenswrapper[4777]: I1124 18:10:01.394754 4777 scope.go:117] "RemoveContainer" containerID="3604715c19080e139ef8fdae4ce5ed671a6ed6d3a8c690fc23fc2f851556f967" Nov 24 18:10:01 crc kubenswrapper[4777]: I1124 18:10:01.551871 4777 generic.go:334] "Generic (PLEG): container finished" podID="a0ef86d3-63fe-421c-b274-eee544924006" containerID="a5a2aca4ed00f1883aa49419845e5ddcf3ef7432e5e6ad690124e5a7fecb7b93" exitCode=0 Nov 24 18:10:01 crc kubenswrapper[4777]: I1124 18:10:01.551927 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w8n5g/must-gather-tvr92" event={"ID":"a0ef86d3-63fe-421c-b274-eee544924006","Type":"ContainerDied","Data":"a5a2aca4ed00f1883aa49419845e5ddcf3ef7432e5e6ad690124e5a7fecb7b93"} Nov 24 18:10:01 crc kubenswrapper[4777]: I1124 18:10:01.552561 4777 scope.go:117] "RemoveContainer" containerID="a5a2aca4ed00f1883aa49419845e5ddcf3ef7432e5e6ad690124e5a7fecb7b93" Nov 24 18:10:02 crc kubenswrapper[4777]: I1124 18:10:02.594717 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w8n5g_must-gather-tvr92_a0ef86d3-63fe-421c-b274-eee544924006/gather/0.log" Nov 24 18:10:14 crc kubenswrapper[4777]: I1124 18:10:14.219038 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-w8n5g/must-gather-tvr92"] Nov 24 18:10:14 crc kubenswrapper[4777]: I1124 18:10:14.219961 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-w8n5g/must-gather-tvr92" podUID="a0ef86d3-63fe-421c-b274-eee544924006" containerName="copy" containerID="cri-o://7232f211d58149ba1bf05b71c2cff6ccd64cc2c20ca455d6d9afb5dc50827621" gracePeriod=2 Nov 24 18:10:14 crc kubenswrapper[4777]: I1124 18:10:14.230089 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-w8n5g/must-gather-tvr92"] Nov 24 18:10:14 crc kubenswrapper[4777]: I1124 18:10:14.704233 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w8n5g_must-gather-tvr92_a0ef86d3-63fe-421c-b274-eee544924006/copy/0.log" Nov 24 18:10:14 crc kubenswrapper[4777]: I1124 18:10:14.706647 4777 generic.go:334] "Generic (PLEG): container finished" podID="a0ef86d3-63fe-421c-b274-eee544924006" containerID="7232f211d58149ba1bf05b71c2cff6ccd64cc2c20ca455d6d9afb5dc50827621" exitCode=143 Nov 24 18:10:14 crc kubenswrapper[4777]: I1124 18:10:14.894473 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w8n5g_must-gather-tvr92_a0ef86d3-63fe-421c-b274-eee544924006/copy/0.log" Nov 24 18:10:14 crc kubenswrapper[4777]: I1124 18:10:14.894924 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/must-gather-tvr92" Nov 24 18:10:14 crc kubenswrapper[4777]: I1124 18:10:14.995844 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a0ef86d3-63fe-421c-b274-eee544924006-must-gather-output\") pod \"a0ef86d3-63fe-421c-b274-eee544924006\" (UID: \"a0ef86d3-63fe-421c-b274-eee544924006\") " Nov 24 18:10:14 crc kubenswrapper[4777]: I1124 18:10:14.995987 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llvdc\" (UniqueName: \"kubernetes.io/projected/a0ef86d3-63fe-421c-b274-eee544924006-kube-api-access-llvdc\") pod \"a0ef86d3-63fe-421c-b274-eee544924006\" (UID: \"a0ef86d3-63fe-421c-b274-eee544924006\") " Nov 24 18:10:15 crc kubenswrapper[4777]: I1124 18:10:15.001446 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0ef86d3-63fe-421c-b274-eee544924006-kube-api-access-llvdc" (OuterVolumeSpecName: "kube-api-access-llvdc") pod "a0ef86d3-63fe-421c-b274-eee544924006" (UID: "a0ef86d3-63fe-421c-b274-eee544924006"). InnerVolumeSpecName "kube-api-access-llvdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:10:15 crc kubenswrapper[4777]: I1124 18:10:15.097916 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llvdc\" (UniqueName: \"kubernetes.io/projected/a0ef86d3-63fe-421c-b274-eee544924006-kube-api-access-llvdc\") on node \"crc\" DevicePath \"\"" Nov 24 18:10:15 crc kubenswrapper[4777]: I1124 18:10:15.152408 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0ef86d3-63fe-421c-b274-eee544924006-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a0ef86d3-63fe-421c-b274-eee544924006" (UID: "a0ef86d3-63fe-421c-b274-eee544924006"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:10:15 crc kubenswrapper[4777]: I1124 18:10:15.199451 4777 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a0ef86d3-63fe-421c-b274-eee544924006-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 18:10:15 crc kubenswrapper[4777]: I1124 18:10:15.258599 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0ef86d3-63fe-421c-b274-eee544924006" path="/var/lib/kubelet/pods/a0ef86d3-63fe-421c-b274-eee544924006/volumes" Nov 24 18:10:15 crc kubenswrapper[4777]: I1124 18:10:15.723307 4777 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w8n5g_must-gather-tvr92_a0ef86d3-63fe-421c-b274-eee544924006/copy/0.log" Nov 24 18:10:15 crc kubenswrapper[4777]: I1124 18:10:15.724146 4777 scope.go:117] "RemoveContainer" containerID="7232f211d58149ba1bf05b71c2cff6ccd64cc2c20ca455d6d9afb5dc50827621" Nov 24 18:10:15 crc kubenswrapper[4777]: I1124 18:10:15.724179 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w8n5g/must-gather-tvr92" Nov 24 18:10:15 crc kubenswrapper[4777]: I1124 18:10:15.760861 4777 scope.go:117] "RemoveContainer" containerID="a5a2aca4ed00f1883aa49419845e5ddcf3ef7432e5e6ad690124e5a7fecb7b93" Nov 24 18:11:01 crc kubenswrapper[4777]: I1124 18:11:01.464542 4777 scope.go:117] "RemoveContainer" containerID="c74f2257af9b0395bece23b5a83fb192cb3088590e1cee4c812edb1b9c71d517" Nov 24 18:11:44 crc kubenswrapper[4777]: I1124 18:11:44.270810 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:11:44 crc kubenswrapper[4777]: I1124 18:11:44.271352 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.630382 4777 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mq5xw"] Nov 24 18:12:05 crc kubenswrapper[4777]: E1124 18:12:05.631665 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7e48980-5467-4d69-9a43-3437ca6aa0b7" containerName="extract-utilities" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.631690 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7e48980-5467-4d69-9a43-3437ca6aa0b7" containerName="extract-utilities" Nov 24 18:12:05 crc kubenswrapper[4777]: E1124 18:12:05.631727 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ef86d3-63fe-421c-b274-eee544924006" containerName="copy" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.631738 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ef86d3-63fe-421c-b274-eee544924006" containerName="copy" Nov 24 18:12:05 crc kubenswrapper[4777]: E1124 18:12:05.631760 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ef86d3-63fe-421c-b274-eee544924006" containerName="gather" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.631770 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ef86d3-63fe-421c-b274-eee544924006" containerName="gather" Nov 24 18:12:05 crc kubenswrapper[4777]: E1124 18:12:05.631795 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7e48980-5467-4d69-9a43-3437ca6aa0b7" containerName="registry-server" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.631806 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7e48980-5467-4d69-9a43-3437ca6aa0b7" containerName="registry-server" Nov 24 18:12:05 crc kubenswrapper[4777]: E1124 18:12:05.631842 4777 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7e48980-5467-4d69-9a43-3437ca6aa0b7" containerName="extract-content" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.631852 4777 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7e48980-5467-4d69-9a43-3437ca6aa0b7" containerName="extract-content" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.632205 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ef86d3-63fe-421c-b274-eee544924006" containerName="copy" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.632246 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7e48980-5467-4d69-9a43-3437ca6aa0b7" containerName="registry-server" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.632269 4777 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ef86d3-63fe-421c-b274-eee544924006" containerName="gather" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.635404 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.662273 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mq5xw"] Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.776927 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-catalog-content\") pod \"redhat-marketplace-mq5xw\" (UID: \"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0\") " pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.777036 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-utilities\") pod \"redhat-marketplace-mq5xw\" (UID: \"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0\") " pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.777674 4777 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9hkm\" (UniqueName: \"kubernetes.io/projected/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-kube-api-access-d9hkm\") pod \"redhat-marketplace-mq5xw\" (UID: \"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0\") " pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.879717 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9hkm\" (UniqueName: \"kubernetes.io/projected/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-kube-api-access-d9hkm\") pod \"redhat-marketplace-mq5xw\" (UID: \"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0\") " pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.879784 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-catalog-content\") pod \"redhat-marketplace-mq5xw\" (UID: \"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0\") " pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.879810 4777 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-utilities\") pod \"redhat-marketplace-mq5xw\" (UID: \"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0\") " pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.880289 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-utilities\") pod \"redhat-marketplace-mq5xw\" (UID: \"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0\") " pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.880415 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-catalog-content\") pod \"redhat-marketplace-mq5xw\" (UID: \"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0\") " pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.951722 4777 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9hkm\" (UniqueName: \"kubernetes.io/projected/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-kube-api-access-d9hkm\") pod \"redhat-marketplace-mq5xw\" (UID: \"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0\") " pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:05 crc kubenswrapper[4777]: I1124 18:12:05.963609 4777 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:06 crc kubenswrapper[4777]: I1124 18:12:06.411719 4777 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mq5xw"] Nov 24 18:12:07 crc kubenswrapper[4777]: I1124 18:12:07.065094 4777 generic.go:334] "Generic (PLEG): container finished" podID="e0d506b2-5b4d-4881-8757-dc9ddc4f76f0" containerID="dda0bb09d6b2a04d666c43a0e8578fbe0d97b69d2455ba66b7f7b7a42aff8fdc" exitCode=0 Nov 24 18:12:07 crc kubenswrapper[4777]: I1124 18:12:07.065143 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mq5xw" event={"ID":"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0","Type":"ContainerDied","Data":"dda0bb09d6b2a04d666c43a0e8578fbe0d97b69d2455ba66b7f7b7a42aff8fdc"} Nov 24 18:12:07 crc kubenswrapper[4777]: I1124 18:12:07.065447 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mq5xw" event={"ID":"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0","Type":"ContainerStarted","Data":"3412ee9f5475d9adbb7e350f5f6d643b16623ca86b955511fc76b85f37c9f661"} Nov 24 18:12:07 crc kubenswrapper[4777]: I1124 18:12:07.067341 4777 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 18:12:08 crc kubenswrapper[4777]: I1124 18:12:08.075391 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mq5xw" event={"ID":"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0","Type":"ContainerStarted","Data":"7c2f4dfc65c0e065d7a697719fa90232c616c50b3ada966d6880a04d810b0852"} Nov 24 18:12:09 crc kubenswrapper[4777]: I1124 18:12:09.087281 4777 generic.go:334] "Generic (PLEG): container finished" podID="e0d506b2-5b4d-4881-8757-dc9ddc4f76f0" containerID="7c2f4dfc65c0e065d7a697719fa90232c616c50b3ada966d6880a04d810b0852" exitCode=0 Nov 24 18:12:09 crc kubenswrapper[4777]: I1124 18:12:09.087326 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mq5xw" event={"ID":"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0","Type":"ContainerDied","Data":"7c2f4dfc65c0e065d7a697719fa90232c616c50b3ada966d6880a04d810b0852"} Nov 24 18:12:10 crc kubenswrapper[4777]: I1124 18:12:10.099359 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mq5xw" event={"ID":"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0","Type":"ContainerStarted","Data":"0a5c040bc8aee2def8730d70b0799d768abd9bf920c3a644468fb3305246d1a5"} Nov 24 18:12:10 crc kubenswrapper[4777]: I1124 18:12:10.121231 4777 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mq5xw" podStartSLOduration=2.7058698420000002 podStartE2EDuration="5.121214605s" podCreationTimestamp="2025-11-24 18:12:05 +0000 UTC" firstStartedPulling="2025-11-24 18:12:07.067119689 +0000 UTC m=+4285.226314738" lastFinishedPulling="2025-11-24 18:12:09.482464442 +0000 UTC m=+4287.641659501" observedRunningTime="2025-11-24 18:12:10.115384782 +0000 UTC m=+4288.274579831" watchObservedRunningTime="2025-11-24 18:12:10.121214605 +0000 UTC m=+4288.280409654" Nov 24 18:12:14 crc kubenswrapper[4777]: I1124 18:12:14.270660 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:12:14 crc kubenswrapper[4777]: I1124 18:12:14.271936 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:12:15 crc kubenswrapper[4777]: I1124 18:12:15.964603 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:15 crc kubenswrapper[4777]: I1124 18:12:15.964957 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:16 crc kubenswrapper[4777]: I1124 18:12:16.008539 4777 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:16 crc kubenswrapper[4777]: I1124 18:12:16.209182 4777 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:16 crc kubenswrapper[4777]: I1124 18:12:16.252630 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mq5xw"] Nov 24 18:12:18 crc kubenswrapper[4777]: I1124 18:12:18.184476 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mq5xw" podUID="e0d506b2-5b4d-4881-8757-dc9ddc4f76f0" containerName="registry-server" containerID="cri-o://0a5c040bc8aee2def8730d70b0799d768abd9bf920c3a644468fb3305246d1a5" gracePeriod=2 Nov 24 18:12:18 crc kubenswrapper[4777]: I1124 18:12:18.905688 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:18 crc kubenswrapper[4777]: I1124 18:12:18.954560 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-catalog-content\") pod \"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0\" (UID: \"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0\") " Nov 24 18:12:18 crc kubenswrapper[4777]: I1124 18:12:18.954636 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9hkm\" (UniqueName: \"kubernetes.io/projected/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-kube-api-access-d9hkm\") pod \"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0\" (UID: \"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0\") " Nov 24 18:12:18 crc kubenswrapper[4777]: I1124 18:12:18.954718 4777 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-utilities\") pod \"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0\" (UID: \"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0\") " Nov 24 18:12:18 crc kubenswrapper[4777]: I1124 18:12:18.955749 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-utilities" (OuterVolumeSpecName: "utilities") pod "e0d506b2-5b4d-4881-8757-dc9ddc4f76f0" (UID: "e0d506b2-5b4d-4881-8757-dc9ddc4f76f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:12:18 crc kubenswrapper[4777]: I1124 18:12:18.960738 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-kube-api-access-d9hkm" (OuterVolumeSpecName: "kube-api-access-d9hkm") pod "e0d506b2-5b4d-4881-8757-dc9ddc4f76f0" (UID: "e0d506b2-5b4d-4881-8757-dc9ddc4f76f0"). InnerVolumeSpecName "kube-api-access-d9hkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 18:12:18 crc kubenswrapper[4777]: I1124 18:12:18.971416 4777 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0d506b2-5b4d-4881-8757-dc9ddc4f76f0" (UID: "e0d506b2-5b4d-4881-8757-dc9ddc4f76f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.057739 4777 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.057767 4777 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.057780 4777 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9hkm\" (UniqueName: \"kubernetes.io/projected/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0-kube-api-access-d9hkm\") on node \"crc\" DevicePath \"\"" Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.199450 4777 generic.go:334] "Generic (PLEG): container finished" podID="e0d506b2-5b4d-4881-8757-dc9ddc4f76f0" containerID="0a5c040bc8aee2def8730d70b0799d768abd9bf920c3a644468fb3305246d1a5" exitCode=0 Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.199555 4777 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mq5xw" Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.199544 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mq5xw" event={"ID":"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0","Type":"ContainerDied","Data":"0a5c040bc8aee2def8730d70b0799d768abd9bf920c3a644468fb3305246d1a5"} Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.199743 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mq5xw" event={"ID":"e0d506b2-5b4d-4881-8757-dc9ddc4f76f0","Type":"ContainerDied","Data":"3412ee9f5475d9adbb7e350f5f6d643b16623ca86b955511fc76b85f37c9f661"} Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.199770 4777 scope.go:117] "RemoveContainer" containerID="0a5c040bc8aee2def8730d70b0799d768abd9bf920c3a644468fb3305246d1a5" Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.220230 4777 scope.go:117] "RemoveContainer" containerID="7c2f4dfc65c0e065d7a697719fa90232c616c50b3ada966d6880a04d810b0852" Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.241205 4777 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mq5xw"] Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.247315 4777 scope.go:117] "RemoveContainer" containerID="dda0bb09d6b2a04d666c43a0e8578fbe0d97b69d2455ba66b7f7b7a42aff8fdc" Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.276798 4777 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mq5xw"] Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.310897 4777 scope.go:117] "RemoveContainer" containerID="0a5c040bc8aee2def8730d70b0799d768abd9bf920c3a644468fb3305246d1a5" Nov 24 18:12:19 crc kubenswrapper[4777]: E1124 18:12:19.311491 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a5c040bc8aee2def8730d70b0799d768abd9bf920c3a644468fb3305246d1a5\": container with ID starting with 0a5c040bc8aee2def8730d70b0799d768abd9bf920c3a644468fb3305246d1a5 not found: ID does not exist" containerID="0a5c040bc8aee2def8730d70b0799d768abd9bf920c3a644468fb3305246d1a5" Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.311542 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a5c040bc8aee2def8730d70b0799d768abd9bf920c3a644468fb3305246d1a5"} err="failed to get container status \"0a5c040bc8aee2def8730d70b0799d768abd9bf920c3a644468fb3305246d1a5\": rpc error: code = NotFound desc = could not find container \"0a5c040bc8aee2def8730d70b0799d768abd9bf920c3a644468fb3305246d1a5\": container with ID starting with 0a5c040bc8aee2def8730d70b0799d768abd9bf920c3a644468fb3305246d1a5 not found: ID does not exist" Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.311570 4777 scope.go:117] "RemoveContainer" containerID="7c2f4dfc65c0e065d7a697719fa90232c616c50b3ada966d6880a04d810b0852" Nov 24 18:12:19 crc kubenswrapper[4777]: E1124 18:12:19.311817 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c2f4dfc65c0e065d7a697719fa90232c616c50b3ada966d6880a04d810b0852\": container with ID starting with 7c2f4dfc65c0e065d7a697719fa90232c616c50b3ada966d6880a04d810b0852 not found: ID does not exist" containerID="7c2f4dfc65c0e065d7a697719fa90232c616c50b3ada966d6880a04d810b0852" Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.311855 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c2f4dfc65c0e065d7a697719fa90232c616c50b3ada966d6880a04d810b0852"} err="failed to get container status \"7c2f4dfc65c0e065d7a697719fa90232c616c50b3ada966d6880a04d810b0852\": rpc error: code = NotFound desc = could not find container \"7c2f4dfc65c0e065d7a697719fa90232c616c50b3ada966d6880a04d810b0852\": container with ID starting with 7c2f4dfc65c0e065d7a697719fa90232c616c50b3ada966d6880a04d810b0852 not found: ID does not exist" Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.311873 4777 scope.go:117] "RemoveContainer" containerID="dda0bb09d6b2a04d666c43a0e8578fbe0d97b69d2455ba66b7f7b7a42aff8fdc" Nov 24 18:12:19 crc kubenswrapper[4777]: E1124 18:12:19.312286 4777 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dda0bb09d6b2a04d666c43a0e8578fbe0d97b69d2455ba66b7f7b7a42aff8fdc\": container with ID starting with dda0bb09d6b2a04d666c43a0e8578fbe0d97b69d2455ba66b7f7b7a42aff8fdc not found: ID does not exist" containerID="dda0bb09d6b2a04d666c43a0e8578fbe0d97b69d2455ba66b7f7b7a42aff8fdc" Nov 24 18:12:19 crc kubenswrapper[4777]: I1124 18:12:19.312313 4777 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda0bb09d6b2a04d666c43a0e8578fbe0d97b69d2455ba66b7f7b7a42aff8fdc"} err="failed to get container status \"dda0bb09d6b2a04d666c43a0e8578fbe0d97b69d2455ba66b7f7b7a42aff8fdc\": rpc error: code = NotFound desc = could not find container \"dda0bb09d6b2a04d666c43a0e8578fbe0d97b69d2455ba66b7f7b7a42aff8fdc\": container with ID starting with dda0bb09d6b2a04d666c43a0e8578fbe0d97b69d2455ba66b7f7b7a42aff8fdc not found: ID does not exist" Nov 24 18:12:21 crc kubenswrapper[4777]: I1124 18:12:21.266753 4777 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0d506b2-5b4d-4881-8757-dc9ddc4f76f0" path="/var/lib/kubelet/pods/e0d506b2-5b4d-4881-8757-dc9ddc4f76f0/volumes" Nov 24 18:12:44 crc kubenswrapper[4777]: I1124 18:12:44.270764 4777 patch_prober.go:28] interesting pod/machine-config-daemon-vtgz9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 18:12:44 crc kubenswrapper[4777]: I1124 18:12:44.271393 4777 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 18:12:44 crc kubenswrapper[4777]: I1124 18:12:44.271442 4777 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" Nov 24 18:12:44 crc kubenswrapper[4777]: I1124 18:12:44.272359 4777 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e73f0ac9e582e0fd5e9016a5582a523e4e3ccdb50f6f1d6e5751489d53ee9df"} pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 18:12:44 crc kubenswrapper[4777]: I1124 18:12:44.272426 4777 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" containerName="machine-config-daemon" containerID="cri-o://8e73f0ac9e582e0fd5e9016a5582a523e4e3ccdb50f6f1d6e5751489d53ee9df" gracePeriod=600 Nov 24 18:12:44 crc kubenswrapper[4777]: E1124 18:12:44.397171 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:12:44 crc kubenswrapper[4777]: I1124 18:12:44.458191 4777 generic.go:334] "Generic (PLEG): container finished" podID="0d953828-f624-4e22-890b-8a3a704547a7" containerID="8e73f0ac9e582e0fd5e9016a5582a523e4e3ccdb50f6f1d6e5751489d53ee9df" exitCode=0 Nov 24 18:12:44 crc kubenswrapper[4777]: I1124 18:12:44.458248 4777 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" event={"ID":"0d953828-f624-4e22-890b-8a3a704547a7","Type":"ContainerDied","Data":"8e73f0ac9e582e0fd5e9016a5582a523e4e3ccdb50f6f1d6e5751489d53ee9df"} Nov 24 18:12:44 crc kubenswrapper[4777]: I1124 18:12:44.458341 4777 scope.go:117] "RemoveContainer" containerID="1b72042dc72b04f0b4e64de35ea4d6163e72725033a38990bed0f8f82baa07f0" Nov 24 18:12:44 crc kubenswrapper[4777]: I1124 18:12:44.459109 4777 scope.go:117] "RemoveContainer" containerID="8e73f0ac9e582e0fd5e9016a5582a523e4e3ccdb50f6f1d6e5751489d53ee9df" Nov 24 18:12:44 crc kubenswrapper[4777]: E1124 18:12:44.459441 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:12:57 crc kubenswrapper[4777]: I1124 18:12:57.245464 4777 scope.go:117] "RemoveContainer" containerID="8e73f0ac9e582e0fd5e9016a5582a523e4e3ccdb50f6f1d6e5751489d53ee9df" Nov 24 18:12:57 crc kubenswrapper[4777]: E1124 18:12:57.246506 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:13:10 crc kubenswrapper[4777]: I1124 18:13:10.247232 4777 scope.go:117] "RemoveContainer" containerID="8e73f0ac9e582e0fd5e9016a5582a523e4e3ccdb50f6f1d6e5751489d53ee9df" Nov 24 18:13:10 crc kubenswrapper[4777]: E1124 18:13:10.248606 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:13:24 crc kubenswrapper[4777]: I1124 18:13:24.246051 4777 scope.go:117] "RemoveContainer" containerID="8e73f0ac9e582e0fd5e9016a5582a523e4e3ccdb50f6f1d6e5751489d53ee9df" Nov 24 18:13:24 crc kubenswrapper[4777]: E1124 18:13:24.247017 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:13:36 crc kubenswrapper[4777]: I1124 18:13:36.245153 4777 scope.go:117] "RemoveContainer" containerID="8e73f0ac9e582e0fd5e9016a5582a523e4e3ccdb50f6f1d6e5751489d53ee9df" Nov 24 18:13:36 crc kubenswrapper[4777]: E1124 18:13:36.246107 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:13:50 crc kubenswrapper[4777]: I1124 18:13:50.245647 4777 scope.go:117] "RemoveContainer" containerID="8e73f0ac9e582e0fd5e9016a5582a523e4e3ccdb50f6f1d6e5751489d53ee9df" Nov 24 18:13:50 crc kubenswrapper[4777]: E1124 18:13:50.246523 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:14:05 crc kubenswrapper[4777]: I1124 18:14:05.245509 4777 scope.go:117] "RemoveContainer" containerID="8e73f0ac9e582e0fd5e9016a5582a523e4e3ccdb50f6f1d6e5751489d53ee9df" Nov 24 18:14:05 crc kubenswrapper[4777]: E1124 18:14:05.246353 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:14:16 crc kubenswrapper[4777]: I1124 18:14:16.246134 4777 scope.go:117] "RemoveContainer" containerID="8e73f0ac9e582e0fd5e9016a5582a523e4e3ccdb50f6f1d6e5751489d53ee9df" Nov 24 18:14:16 crc kubenswrapper[4777]: E1124 18:14:16.246890 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" Nov 24 18:14:28 crc kubenswrapper[4777]: I1124 18:14:28.244916 4777 scope.go:117] "RemoveContainer" containerID="8e73f0ac9e582e0fd5e9016a5582a523e4e3ccdb50f6f1d6e5751489d53ee9df" Nov 24 18:14:28 crc kubenswrapper[4777]: E1124 18:14:28.245794 4777 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vtgz9_openshift-machine-config-operator(0d953828-f624-4e22-890b-8a3a704547a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-vtgz9" podUID="0d953828-f624-4e22-890b-8a3a704547a7" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111120225024433 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111120226017351 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111107152016477 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111107152015447 5ustar corecore